Parker Dunlap 11/15/2013 Semantic image analysis techniques can automatically detect high level content of images Lack of intuitive visualization and analysis techniques Allow users to effectively browse/search in large databases Allow analysts to evaluate their annotation process through interactive visual exploration Target search ◦ User knows exactly what they want, a precise image Search by association ◦ Find interesting things related to certain image Category search ◦ Retrieve images that are representative of a certain class Semantic contents of images are more useful for image exploration than low level features but in most large scale image collections (internet) semantics are usually not described This has given rise to techniques that enable automatic annotation of images according to their semantic concepts Contains semantic image classification process that automatically annotates large image collections Contains coordinated visualization techniques that allow interactive exploration Contains visualization techniques that allow analysts to evaluate and monitor annotation process 1. 2. 3. Annotation Engine Image Browsing Interface Visual Image Analysis Abstract Image content by detecting underlying salient objects (distinguishable regions) Associate salient objects with corresponding semantic objects according to their perceptual properties Keywords for semantic objects are used to annotate the image Highlighted regions are salient objects detected and associated with semantic object “sand field” Goal is to bridge the gap between low-level visual features and high-level semantic concepts Annotation engine has set of predefined salient objects and functions to detect them from images ◦ Uses techniques like image segmentation and SVM classifiers Annotation engine assigns a semantic concept to the data based on semantic content ◦ Sand, Field, Water → Seaworld ◦ Flowers, Trees → Garden Image overview using MDS ◦ Use the annotations to calculate distance matrix and input into MDS algorithm Distance between each pair of images in the content space ◦ Algorithm outputs a 2D position for each image based on similarity with other images Maps image miniatures onto the screen based on their content similarities Similar images placed closer to each other Goal of MDS is to map some high dimensional data into lower dimension (in our case 2D) ◦ To learn more about MDS see MDS Overview Visually represents the contents of the entire collection of images Correlations of different contents and detailed annotations are displayed Interactively exploring large datasets with real time response (high dimensionality) Block of pixels to represent images contents Each image is mapped to a pixel whose color indicates if the image contains/doesn’t contain the content for that block Pixel representing the same image is the same for all blocks Allows us to observe content of image collection by scanning labels of the blocks Can see correlations among the contents Can also select images to see them highlighted in the view Position of the blocks are determined by similarity with neighboring contents Pixels are generally created in a spiral arrangement starting from the center and moving out Pixel order can greatly effect the looks of VaR view To increase scalability, interface users miniature versions of images ◦ High res original pictures would increase load times Load image miniatures as textures objects in OpenGL ◦ Allows all interactions to be done in real time To reduce clutter in the MDS overview, the system provides many interactions ◦ ◦ ◦ ◦ ◦ ◦ ◦ Reordering Dynamic Scaling Relocation Distortion Showing Original Image Zoom Pan Reordering ◦ Randomizing order of all images allows each frame to have an equal probability of being visible ◦ User can also explicitly bring certain image to the front by selecting it Dynamic Scaling ◦ Interactively reduce image miniature size to reduce overlap or increase image size to examine detail Relocation ◦ Manually change position of individual image by dragging and dropping Distortion ◦ Enlarge size of certain image(s) while retaining size of all others Showing Original Image ◦ Actual image (instead of scaled down image used by OpenGL) opens at full resolution in new window ◦ Only loaded when requested to save space/time Zoom/Pan ◦ Zoom in/out and pan left/right Can use multiple techniques at once to achieve some goal ◦ Use Dynamic Scaling with zooming in to examine local details with less clutter Selection ◦ Interactively select a sample image to see similar images in display ◦ Can change similarity threshold via a slider to increase/decrease number of results Sorting ◦ Images can be sorted by concepts or similarity to selected image Inspired by rainfall animation Correlations between image of interest and other images are modeled through an animation Focus image is on the bottom (ground) and the other images fall to the ground (rain) at accelerations related to their similarity Search for images with/without certain content Reduce a selected subset by requiring images must/not contain certain content Increase selected subset by adding new images All these functions done by clicking on images while holding certain function key Offers many similar interactions as MDS as well Each image has its visual representations in both MDS and VaR views Selected images are highlighted in both views Can use appropriate view as needed ◦ MDS to select image based on relationship to sample image ◦ VaR to select image based on content Common strategy is to start from VaR and switch to MDS after number of images has been greatly reduced We can use the MDS and VaR views to see how well the annotations of images correspond to their actual content Select “red-flower” images from VaR view and verify using MDS view to see if the images are actually red flowers If automatic annotation makes a mistake, user can manually annotate image to fix it VaR display also shows the reliability of the annotation by surrounding it with a colored frame ◦ Green is safe to use, Yellow means lower reliability measure Reliability measure can be determined from annotation process or manually set up by analysts Comparison of SIB to the sequential thumbnail view from Microsoft Explorer Modes used in Microsoft Explorer ◦ Random Explorer – images are randomly sorted ◦ Sorted Explorer – images are sorted according to semantic concepts generated by the classification process 10 participants from varying fields Each subject used both Sorted Explorer and SIB ◦ Random Explorer was only tested on 3 participants since expected results were so low Participants given 3 tasks to perform on 2 data sets ◦ 180 second timeout window 1. 2. 3. Presented with a particular image and asked to search for it from the 1100 images in the data set Asked to find images containing particular features (sand, water, sky, etc…) Asked to approximate what proportion of the images in the dataset contained particular contents (% that contain mountains) Random Explorer ◦ 2/9 trials failed ◦ 81 seconds was average time with 29 seconds standard deviation Sorted Explorer SIB ◦ 2/30 trials failed ◦ 29 seconds was average time with 20 seconds standard deviation ◦ 6/30 trials failed ◦ 45 seconds was average time with 26 seconds standard deviation Failure in SIB was due to inaccuracy in the annotation process SIB tended to be slower than Sorted Explorer because content names could be confusing ◦ This advantage will decrease as the data set grows because Explorer provides no overview model Task 2 had similar results to Task 1 Task 3 was where SIB became dominant Positive feedback for SIB Enjoyed Search by content feature the most Enjoyed MDS overview over Windows explorer to see entire collection of images at once Suggested side-by-side views, example image next to blocks in VaR view Semantic Image Browser was introduced that attempts to bride information visualization with automatic image annotation MDS image layout that groups images based on semantic similarities VaR content display to represent large image collections Semantic Image Browser: Bridging Information Visualization with Automated Intelligent Image Analysis Value and Relation Display for Interactive Exploration of High Dimensional Datasets MDS Overview