Change search
Refine search result
1 - 15 of 15
CiteExportLink to result list
Permanent link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Rows per page
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sort
  • Standard (Relevance)
  • Author A-Ö
  • Author Ö-A
  • Title A-Ö
  • Title Ö-A
  • Publication type A-Ö
  • Publication type Ö-A
  • Issued (Oldest first)
  • Issued (Newest first)
  • Created (Oldest first)
  • Created (Newest first)
  • Last updated (Oldest first)
  • Last updated (Newest first)
  • Disputation date (earliest first)
  • Disputation date (latest first)
  • Standard (Relevance)
  • Author A-Ö
  • Author Ö-A
  • Title A-Ö
  • Title Ö-A
  • Publication type A-Ö
  • Publication type Ö-A
  • Issued (Oldest first)
  • Issued (Newest first)
  • Created (Oldest first)
  • Created (Newest first)
  • Last updated (Oldest first)
  • Last updated (Newest first)
  • Disputation date (earliest first)
  • Disputation date (latest first)
Select
The maximal number of hits you can export is 250. When you want to export more records please use the Create feeds function.
  • 1.
    Abedan Kondori, Farid
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Yousefi, Shahrouz
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Smart Baggage in Aviation2011In: 2011 IEEE International Conferences on Internet of Things, and Cyber, Physical and Social Computing / [ed] Feng Xia, Zhikui Chen, Gang Pan, Laurence T. Yang, and Jianhua Ma, Los Alamitos: IEEE Computer Society, 2011, p. 620-623Conference paper (Refereed)
    Abstract [en]

    Nowadays, the Internet has dramatically changed the way people take the normal course of actions. By the recent growth of the Internet, connecting different objects to users through mobile phones and computers is no longer a dream. Aviation industry is one of the areas which have a strong potential to benefit from the Internet of Things. Among many problems related to air travel, delayed and lost luggage are the most common and irritating. Therefore, this paper suggests anew baggage control system, where users can simply track their baggage at the airport to avoid losing them. Attaching a particular pattern on the bag, which can be detected and localized from long distance by an ordinary camera, users are able to track their baggage. The proposed system is much cheaper than previous implementations and does not require sophisticated equipment.

  • 2.
    Abedan Kondori, Farid
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Yousefi, Shahrouz
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Direct three-dimensional head pose estimation from Kinect-type sensors2014In: Electronics Letters, ISSN 0013-5194, E-ISSN 1350-911X, Vol. 50, no 4, p. 268-270Article in journal (Refereed)
    Abstract [en]

    A direct method for recovering three-dimensional (3D) head motion parameters from a sequence of range images acquired by Kinect sensors is presented. Based on the range images, a new version of the optical flow constraint equation is derived, which can be used to directly estimate 3D motion parameters without any need of imposing other constraints. Since all calculations with the new constraint equation are based on the range images, Z(xyt), the existing techniques and experiences developed and accumulated on the topic of motion from optical flow can be directly applied simply by treating the range images as normal intensity images I(xyt). In this reported work, it is demonstrated how to employ the new optical flow constraint equation to recover the 3D motion of a moving head from the sequences of range images, and furthermore, how to use an old trick to handle the case when the optical flow is large. It is shown, in the end, that the performance of the proposed approach is comparable with that of some of the state-of-the-art approaches that use range data to recover 3D motion parameters.

  • 3.
    Abedan Kondori, Farid
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Yousefi, Shahrouz
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Gesture Tracking for 3D Interaction in Augmented Environments2011Conference paper (Other academic)
  • 4.
    Abedan Kondori, Farid
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Yousefi, Shahrouz
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Real 3D Interaction Behind Mobile Phones for Augmented Environments2011In: 2011 IEEE International Conference on Multimedia and Expo (ICME), IEEE conference proceedings, 2011, p. 1-6Conference paper (Refereed)
    Abstract [en]

    Number of mobile devices such as mobile phones or PDAs has been dramatically increased over the recent years. New mobile devices are equipped with integrated cameras and large displays which make the interaction with device easier and more efficient. Although most of the previous works on interaction between humans and mobile devices are based on 2D touch-screen displays, camera-based interaction opens a new way to manipulate in 3D space behind the device in the camera's field of view. This paper suggests the use of particular patterns from local orientation of the image called Rotational Symmetries to detect and localize human gesture. Relative rotation and translation of human gesture between consecutive frames are estimated by means of extracting stable features. Consequently, this information can be used to facilitate the 3D manipulation of virtual objects in various applications in mobile devices.

  • 5.
    Abedan Kondori, Farid
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Yousefi, Shahrouz
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Sonning, Samuel
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Sonning, Sabina
    3D Head Pose Estimation Using the Kinect2011Conference paper (Refereed)
    Abstract [en]

    Head pose estimation plays an essential role for bridging the information gap between humans and computers. Conventional head pose estimation methods are mostly done in images captured by cameras. However accurate and robust pose estimation is often problematic. In this paper we present an algorithm for recovering the six degrees of freedom (DOF) of motion of a head from a sequence of range images taken by the Microsoft Kinectfor Xbox 360. The proposed algorithm utilizes a least-squares minimization of the difference between themeasured rate of change of depth at a point and the rate predicted by the depth rate constraint equation. We segment the human head from its surroundings and background, and then we estimate the head motion. Our system has the capability to recover the six DOF of the head motion of multiple people in one image. Theproposed system is evaluated in our lab and presents superior results.

  • 6.
    Abedan Kondori, Farid
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Yousefi, Shahrouz
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Liu, Li
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Active human gesture capture for diagnosing and treating movement disorders2013Conference paper (Other academic)
    Abstract [en]

    Movement disorders prevent many people fromenjoying their daily lives. As with other diseases, diagnosisand analysis are key issues in treating such disorders.Computer vision-based motion capture systems are helpfultools for accomplishing this task. However Classical motiontracking systems suffer from several limitations. First theyare not cost effective. Second these systems cannot detectminute motions accurately. Finally they are spatially limitedto the lab environment where the system is installed. In thisproject, we propose an innovative solution to solve the abovementionedissues. Mounting the camera on human body, webuild a convenient, low cost motion capture system that canbe used by the patient in daily-life activities. We refer tothis system as active motion capture, which is not confinedto the lab environment. Real-time experiments in our labrevealed the robustness and accuracy of the system.

  • 7.
    Yousefi, Shahrouz
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    3D photo browsing for future mobile devices2012In: MM 2012 - Proceedings of the 20th ACM International Conference on Multimedia, 2012, p. 1401-1404Conference paper (Refereed)
    Abstract [en]

    By introducing the interactive 3D photo/video browsing and exploration system, we propose novel approaches for handling the limitations of the current 2D mobile technology from two aspects: interaction design and visualization. Our contributions feature an effective interaction that happens in the 3D space behind the mobile device's camera. 3D motion analysis of the user's gesture captured by the device's camera is performed to facilitate the interaction between users and multimedia collections in various applications. This approach will solve a wide range of problems with the current input facilities such as miniature keyboards, tiny joysticks and 2D touch screens. The suggested interactive technology enables users to control, manipulate, organize, and re-arrange their photo/video collections in 3D space using bare-hand, marker-less gesture. Moreover, with the proposed techniques we aim to visualize the 2D photo collection, in 3D, on normal 2D displays. This process is automatically done by retrieving the 3D structure from single images, finding the stereo/multiple views of a scene or using the geo-tagged meta-data from huge photo collections. By using the design and implementation of the contributions of this work, we aim to achieve the following goals: Solving the limitations of the current 2D interaction facilities by 3D gestural interaction; Increasing the usability of the multimedia applications on mobile devices; Enhancing the quality of user experience with the digital collections.

  • 8.
    Yousefi, Shahrouz
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Abedan Kondori, Farid
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Tracking fingers in 3D space for mobile interaction2010Conference paper (Refereed)
    Abstract [en]

    Number of mobile devices such as mobile phones or PDAs has been dramatically increased over the recent years. New mobile devices are equipped with integrated cameras and large displays which make the interaction with device easier and more efficient. Although most of the previous works on interaction between humans and mobile devices are based on 2D touch-screen displays, camera-based interaction opens a new way to manipulate in 3D space behind the device in the camera's field of view.In this paper, our gestural interaction relies heavily on particular patterns from local orientation of the image called Rotational Symmetries. This approach is based on finding the most suitable pattern from the large set of rotational symmetries of different orders which ensures a reliable detector for fingertips and human gesture. Consequently, gesture detection and tracking can be used as an efficient tool for 3D manipulation in various applications in computer vision and augmented reality.

  • 9.
    Yousefi, Shahrouz
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Kondori, Farid Abedan
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    3D Gestural Interaction for Stereoscopic Visualization on Mobile Devices2011In: Computer Analysis of Images and Patterns: 14th International Conference, CAIP 2011, PT 2 / [ed] Real, P; DiazPernil, D; MolinaAbril, H; Berciano, A; Kropatsch, W, Berlin: Springer Berlin/Heidelberg, 2011, p. 555-562Conference paper (Refereed)
    Abstract [en]

    Number of mobile devices such as smart phones or Tablet PCs has been dramatically increased over the recent years. New mobile devices are equipped with integrated cameras and large displays which make the interaction with device more efficient. Although most of the previous works on interaction between humans and mobile devices are based on 2D touch-screen displays, camera-based interaction opens a new way to manipulate in 3D space behind the device in the camera's field of view. In this paper, our gestural interaction heavily relies on particular patterns from local orientation of the image called Rotational Symmetries. This approach is based on finding the most suitable pattern from a large set of rotational symmetries of different orders which ensures a reliable detector for hand gesture. Consequently, gesture detection and tracking can be hired as an efficient tool for 3D manipulation in various applications in computer vision and augmented reality. The final output will be rendered into color anaglyphs for 3D visualization. Depending on the coding technology different low cost 3D glasses will be used for viewers.

  • 10.
    Yousefi, Shahrouz
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Kondori, Farid Abedan
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    3D Visualization of Monocular Images in Photo Collections2011Conference paper (Refereed)
  • 11.
    Yousefi, Shahrouz
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Kondori, Farid Abedan
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    3D visualization of single images using patch level depth2011In: 2011 PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND MULTIMEDIA APPLICATIONS (SIGMAP 2011) / [ed] Barranco, AL & Tsihrintzis, G, IEEE, 2011, p. 1-6Conference paper (Refereed)
    Abstract [en]

    In this paper we consider the task of 3D photo visualization using a single monocular image. The main idea is to use single photos taken by capturing devices such as ordinary cameras, mobile phones, tablet PCs etc. and visualize them in 3D on normal displays. Supervised learning approach is hired to retrieve depth information from single images. This algorithm is based on the hierarchical multi-scale Markov Random Field (MRF) which models the depth based on the multi-scale global and local features and relation between them in a monocular image. Consequently, the estimated depth image is used to allocate the specified depth parameters for each pixel in the 3D map. Accordingly, the multi-level depth adjustments and coding for color anaglyphs is performed. Our system receives a single 2D image as input and provides a anaglyph coded 3D image in output. Depending on the coding technology the special low-cost anaglyph glasses for viewers will be used.

  • 12.
    Yousefi, Shahrouz
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Kondori, Farid Abedan
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Camera-based gesture tracking for 3D interaction behind mobile devices2012In: International journal of pattern recognition and artificial intelligence, ISSN 0218-0014, Vol. 26, no 8, p. 1260008-Article in journal (Refereed)
    Abstract [en]

    Number of mobile devices such as Smartphones or Tablet PCs has been dramatically increased over the recent years. New mobile devices are equipped with integrated cameras and large displays that make the interaction with the device easier and more efficient. Although most of the previous works on interaction between humans and mobile devices are based on 2D touch-screen displays, camera-based interaction opens a new way to manipulate in 3D space behind the device in the camera's field of view. In this paper, our gestural interaction relies on particular patterns from local orientation of the image called rotational symmetries. This approach is based on finding the most suitable pattern from a large set of rotational symmetries of diffrerent orders that ensures a reliable detector for fingertips and user's gesture. Consequently, gesture detection and tracking can be used as an efficient tool for 3D manipulation in various virtual/augmented reality applications.

  • 13.
    Yousefi, Shahrouz
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Kondori, Farid Abedan
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Experiencing real 3D gestural interaction with mobile devices2013In: Pattern Recognition Letters, ISSN 0167-8655, E-ISSN 1872-7344, Vol. 34, no 8, p. 912-921Article in journal (Refereed)
    Abstract [en]

    Number of mobile devices such as smart phones or Tablet PCs has been dramatically increased over the recent years. New mobile devices are equipped with integrated cameras and large displays which make the interaction with the device more efficient. Although most of the previous works on interaction between humans and mobile devices are based on 2D touch-screen displays, camera-based interaction opens a new way to manipulate in 3D space behind the device, in the camera's field of view. In this paper, our gestural interaction heavily relies on particular patterns from local orientation of the image called Rotational Symmetries. This approach is based on finding the most suitable pattern from a large set of rotational symmetries of different orders that ensures a reliable detector for hand gesture. Consequently, gesture detection and tracking can be hired as an efficient tool for 3D manipulation in various applications in computer vision and augmented reality. The final output will be rendered into color anaglyphs for 3D visualization. Depending on the coding technology, different low cost 3D glasses can be used for the viewers. (C) 2013 Elsevier B.V. All rights reserved.

  • 14.
    Yousefi, Shahrouz
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Kondori, Farid Abedan
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Robust correction of 3D geo-metadata in photo collections by forming a photo grid2011In: 2011 International Conference on Wireless Communications and Signal Processing (WCSP), IEEE, 2011, p. 1-5Conference paper (Refereed)
    Abstract [en]

    In this work, we present a technique for efficient and robust estimation of the exact location and orientation of a photo capture device in a large data set. The provided data set includes a set of photos and the associated information from GPS and orientation sensor. This attached metadata is noisy and lacks precision. Our strategy to correct this uncertain data is based on the data fusion between measurement model, derived from sensor data, and signal model given by the computer vision algorithms. Based on the retrieved information from multiple views of a scene we make a grid of images. Our robust feature detection and matching between images result in finding a reliable transformation. Consequently, relative location and orientation of the data set construct the signal model. On the other hand, information extracted from the single images combined with the measurement data make the measurement model. Finally, Kalman filter is used to fuse these two models iteratively and enhance the estimation of the ground truth(GT) location and orientation. Practically, this approach can help us to design a photo browsing system from a huge collection of photos, enabling 3D navigation and exploration of our huge data set.

  • 15.
    Yousefi, Shahrouz
    et al.
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Kondori, Farid Abedan
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Li, Haibo
    Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
    Stereoscopic visualization of monocular images in photo collections2011Conference paper (Refereed)
1 - 15 of 15
CiteExportLink to result list
Permanent link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf