default search action
4th SIGGRAPH Asia 2011: Hong Kong - Posters
- Zhigeng Pan:
SIGGRAPH Asia 2011 Posters, Hong Kong, December 12-15, 2011. ACM 2011, ISBN 978-1-4503-1137-3
Animation
- Daisuke Mima, Hiroyuki Kubo, Akinobu Maejima, Shigeo Morishima:
Automatic generation of facial wrinkles according to expression changes. 1:1 - Antonella Guidazzoli, Luigi Calori, Francesca Delli Ponti, Tiziano Diamanti, Silvano Imboden, Massimo Alessio Mauri, A. Negri, Giosue Boetto Cohen, Sofia Pescarin, Maria Chiara Liguori:
Apa the Etruscan and 2700 years of 3D Bologna history. 2:1 - Adso Fernández-Baena, David Miralles:
Progressive transitions using body part motion graphs. 3:1-3:2 - Tomohiko Mukai:
Modeling ranges of limb motion for real-time inverse kinematics. 4:1 - Katsumi Sato, Yoko Usui, Takaaki Kaiga, Shinichi Watabe:
The evaluation of 3D stereoscopic vision for learning "Kagra". 6:1
Art & design
- Alain Lioret:
Artificial life art composition for cinema: cinema beings. 7:1 - Kenta Mizui, Makoto Okabe, Rikio Onai:
Choreographing amateur performers using video examples. 8:1-8:2 - Hyoyoung Kim, Dongseop Lee, Jin Wan Park:
Textual visualization based on readability. 9:1 - Olga Sourina, Yisi Liu, Minh Khoa Nguyen:
Emotion-enabled EEG-based interaction. 10:1 - Noriyuki Yamamoto, Nahomi Maki, Akihiko Shirai:
Skeleton-based diverse creature design tool for mass production. 11:1 - Sachiko Kodama, Toshiki Abe:
Blooming space: color-changing sculpture that reflects real flower's color sensed by vase-shaped device having embedded full-color sensors and LEDs. 12:1
Image & video processing
- Jean-Charles Bazin, Alexandre Richard, Yu-Wing Tai, Inso Kweon:
Around the world in 80 seconds. 14:1 - Asli Gençtav, Ahmet Oguz Akyüz:
Evaluation of radiometric camera response recovery methods. 15:1 - Chia-hang Ho, Chun-Te Wu, Wei-Jia Huang:
A video depth refinement with circuit model constraint. 16:1 - Ji-yong Kwon, In-Kwon Lee:
Video composition using geodesic distance. 17:1 - Tomoya Hara, Akinobu Maejima, Shigeo Morishima:
Automatic 3D face generation from video with sparse point constraint and dense deformable model. 18:1 - Min Ki Park, Ji-Ho Cho, In Yeop Jang, Seung Joo Lee, Kwan H. Lee:
An iterative joint bilateral filtering for depth refinement of a 3D model. 19:1 - Manhee Lee, Won-Ki Jeong, Hanspeter Pfister:
Interactive large-scale image editing using operator reduction. 20:1
Interaction
- Sheng-Jie Luo, Shih-Wen Huang, En-Hsu Yen, Li-Ting Huang, Bing-Yu Chen, Kwan-Liu Ma:
Presenting time-evolving activities using communication archive data. 21:1 - Seunghyun Woo, Sanghyun Eim, Changmok Kim, Hyojin Suh:
Golmoo: providing a seamless interaction between various devices. 22:1 - Thitirat Siriborvornratanakul, Masanori Sugimoto:
Cooperative personal projection using multiple DLP projectors. 23:1 - Álvaro Cassinelli, Yuko Zhou, Alexis Zerroug, Masatoshi Ishikawa:
The Laser Aura: a prosthesis for emotional expression. 24:1 - Mariza Dima, Mark Wright, John Lee:
A haptically-augmented interface for digital character animation. 25:1 - Jianfeng Xu, Koichi Takagi, Shigeyuki Sakazawa:
Interactive generation of dancing animation with music synchronization. 26:1 - Guangxia Li, Andrew Bragdon, Zhigeng Pan, Mingmin Zhang, Sharon M. Swartz, David H. Laidlaw, Chaoyang Zhang, Hanyu Liu, Jian Chen:
VisBubbles: a workflow-driven framework for scientific data analysis of time-varying biological datasets. 27:1 - Liming Xu, Jeremy Lyle, Yubao Wu, Zhigeng Pan, Mingmin Zhang, David H. Laidlaw, Robert L. Hester, Jian Chen:
HumMod explorer: a multi-scale time-varying human modeling navigator. 28:1
Modeling
- Alice Pui Lam Hung, Tim Wu, Peter J. Hunter, Kumar Mithraratne:
Simulating facial expressions using anatomically accurate biomechanical model. 29:1 - Natapon Pantuwong, Masanori Sugimoto:
A fully automatic rigging algorithm for 3D character animation. 30:1 - Yukari Nakamura, Ayaka Kaneko, Takayuki Itoh:
An accelerated pocket extraction and evaluation technique for druggability analysis with protein surfaces. 31:1 - Byungjoo Kim, Ku-Jin Kim, Ji-Hoon Choi, Nakhoon Baek, Joon-Kyung Seong, Yoo-Joo Choi:
Finding surface atoms of a protein molecule on a GPU. 32:1 - Kangying Cai, Wenfei Jiang, Teng Ma, Jiang Tian, Wencheng Wang, Tao Luo:
Probability model-adaptive coding of point clouds with octree decomposition. 33:1 - Mihayu Kurokawa, Mami Inomata, Takayuki Itoh, Naruhito Toyoda, Hiromi Sasamoto:
Micro-geometric modeling of human face skis for cosmetic analysis. 34:1 - Michele Bousquet, Jonathan Lloyd:
Standards for 3D models: case study. 35a:1 - Wan-Chun Ma, Yi-Hua Tina Wang, Graham Fyffe, Jernej Barbic, Bing-Yu Chen, Paul E. Debevec:
A blendshape model that incorporates physical interaction. 35:1
Production & visual effects
- Takeshi Miura, Takaaki Kaiga, Naho Matsumoto, Hiroaki Katsura, Katsubumi Tajima, Hideo Tamamoto:
Application of the Bayesian information criterion to keyframe extraction from motion capture data. 36:1 - Zhuo Su, Xiaonan Luo, Zhengjie Deng:
Local oscillation suppression based on joint bilateral filtering framework. 37:1 - Younghee Kim, Chung Hwan Lee, Soon Hyoung Pyo, Ki-Hong Kim:
A scalable parallel SPH implementation for distributed computing. 38:1 - Wei-Chin Lin, Wei-Kai Liao, Chao-Hua Lee:
Simulating and rendering wet hair. 39:1 - Hideki Todo, Ken Anjyo:
Hybrid framework for blendshape manipulations. 40:1 - Sun-Young Lee, Yong-Ho Lee, Jin-Wook Beak, Young-Suk Kim, Hye-Rin Kim, In-Kwon Lee:
Video word balloon authoring system. 41:1-41:2 - Francesco Banterle, Matteo Dellepiane, Roberto Scopigno:
High dynamic range images for enhancing low dynamic range content. 42:1 - Tanchao Zhu, Weiming Wang, Yongming Xie, Pheng-Ann Heng:
An ellipsoid-based perspective projection correction for wide-angle images. 43:1
Rendering
- Takafumi Kojima, Ryota Takeuchi, Soji Yamakawa, Taichi Watanabe, Koji Mikami:
Visual simulation of aurora movement. 44:1 - Takashi Matsuo, Koji Mikami, Taichi Watanabe, Kunio Kondo:
Shape oriented line drawing in real-time 3DCG. 45:1 - Takashi Nishiyama, Hayato Ichino, Tsuyoshi Nakamura, Masayoshi Kanoh, Koji Yamada:
Impressionist painterly style transfer based on texture localization. 46:1 - Laurent Belcour, Cyril Soler:
Frequency based kernel estimation for progressive photon mapping. 47:1 - Tomohito Hattori, Hiroyuki Kubo, Shigeo Morishima:
Real time ambient occlusion by curvature dependent occlusion function. 48:1 - Anthony Pajot, Loïc Barthe, Mathias Paulin:
Robust adaptive sampling for Monte-Carlo-based rendering. 49:1 - Alessandro Artusi, Despina Michael, Benjamin Roch, Yiorgos Chrysanthou, Alan Chalmers:
A selective approach for tone mapping high dynamic range content. 50:1 - Lissy Verma, Subodh Kumar:
Illustrative rendering of historical artifacts. 51:1
Virtual & augmented reality
- Koki Nagano, Takeru Utsugi, Kazuhisa Yanaka, Akihiko Shirai, Masayuki Nakajima:
ScritterHDR: multiplex-hidden imaging on high dynamic range projection. 52:1 - Byung-Kuk Seo, Hideaki Uchiyama, Jong-Il Park:
stAR: visualizing constellations with star retrieval. 53:1 - Wee Teck Fong, Cher Jingting, Farzam Farbiz, Zhiyong Huang:
Variable frequency 60-g haptic renderer for virtual tennis simulation. 54:1 - Yoshiki Oka, Tomoaki Kida, Masanobu Yamamoto:
A cartoon-character costume with facial expression. 55:1 - Benjamin Bolte, Gerd Bruder, Frank Steinicke:
Jumping through immersive video games. 56:1 - Ann McNamara, Frederic I. Parke, Mat Sanford:
Exploring the effect of tiling on large displays. 57:1 - Tansel Halic, Woojin Ahn, Suvranu De:
A framework for 3D interactive applications on the web. 58:1
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.