A website that presents the collection through gorgeous visuals is now considered a must for any self-respecting museum. Photographs of objects, of exhibitions and of the museum itself are increasingly frequent interfaces, linking museums, visitors, experts, collections. How are users engaged by these interfaces? Which skills and strategies are needed for this engagement? What are the consequences of visually mediated interfaces for users of digital knowledge in/about/from museums, archives, and other collections? These developments are discussed in terms of their consequences for how museums view their role in a recent article written with Sarah de Rijcke, Image as Interface: consequences for users of museum knowledge. It appears in a special issue of the journal Library Trends on ‘Involving Users in the Co-Construction of Digital Knowledge in Libraries, Archives, and Museums.’
Tag Archives: database
A new application for high end mobile phones called Layar features the real estate database Funda (one of our 4 cases). By panning one’s mobile phone using the built-in camera, this application displays information on objects encountered. For example, by pointing the phone towards a row of houses, any property found in Funda will be highlighted on the screen, and the database entry further consulted.This site has a demo of Layar, which Sara Kjellberg kindly brought to our attention. Layars is currently directed to the Netherlands, and the Funda application will be launched at the beginning of July.
As the name hints, one can layer information onto the images on the mobile phone, so that one’s presence at a geographical point in a space becomes the interface to the database (together with the use of the right mobile phone and subscription!). If I ever get my hands on this, I’ll be very curious to see how the linking of objects viewed through the camera and those in the database in done. I can imagine that GIS information can be used to link the location of the phone and that this information can be used to draw information from the database on the general location. But this is rather coarse. Does image recognition enter the picture to refine the selection of information to be presented? At first blush, this seems like an overly complex computational challenge to be performed on the fly on a mobile phone. So how does it work?