3:I[5613,[],""] 4:I[1778,[],""] 5:I[8094,["749","static/chunks/749-849ff4ad7efada95.js","356","static/chunks/356-26b4e8bcef98a226.js","185","static/chunks/app/layout-3442ea4404468775.js"],"Header"] 0:["gniBkws-MlQwk6TUWY9J-",[[["",{"children":["publications",{"children":["__PAGE__",{}]}]},"$undefined","$undefined",true],["",{"children":["publications",{"children":["__PAGE__",{},["$L1","$L2",null]]},["$","$L3",null,{"parallelRouterKey":"children","segmentPath":["children","publications","children"],"loading":"$undefined","loadingStyles":"$undefined","loadingScripts":"$undefined","hasLoading":false,"error":"$undefined","errorStyles":"$undefined","errorScripts":"$undefined","template":["$","$L4",null,{}],"templateStyles":"$undefined","templateScripts":"$undefined","notFound":"$undefined","notFoundStyles":"$undefined","styles":null}]]},[null,["$","html",null,{"lang":"en","children":[["$","head",null,{"children":[["$","link",null,{"rel":"icon","type":"image/png","sizes":"64x64","href":"/favicon/favicon.ico"}],["$","link",null,{"rel":"mask-icon","href":"/favicon/icon.svg","color":"#000000"}],["$","link",null,{"rel":"shortcut icon","href":"/favicon/favicon.ico"}],["$","meta",null,{"name":"msapplication-TileColor","content":"#000000"}],["$","meta",null,{"name":"theme-color","content":"#fff"}]]}],["$","body",null,{"className":"__className_4ad370 relative","children":[["$","div",null,{"className":"fixed h-screen w-full bg-gradient-to-br from-indigo-50 via-white to-cyan-100"}],["$","$L5",null,{}],["$","main",null,{"className":"flex min-h-screen w-full flex-col items-center justify-center py-10 absolute top-0","children":["$","$L3",null,{"parallelRouterKey":"children","segmentPath":["children"],"loading":"$undefined","loadingStyles":"$undefined","loadingScripts":"$undefined","hasLoading":false,"error":"$undefined","errorStyles":"$undefined","errorScripts":"$undefined","template":["$","$L4",null,{}],"templateStyles":"$undefined","templateScripts":"$undefined","notFound":[["$","title",null,{"children":"404: This page could not be found."}],["$","div",null,{"style":{"fontFamily":"system-ui,\"Segoe UI\",Roboto,Helvetica,Arial,sans-serif,\"Apple Color Emoji\",\"Segoe UI Emoji\"","height":"100vh","textAlign":"center","display":"flex","flexDirection":"column","alignItems":"center","justifyContent":"center"},"children":["$","div",null,{"children":[["$","style",null,{"dangerouslySetInnerHTML":{"__html":"body{color:#000;background:#fff;margin:0}.next-error-h1{border-right:1px solid rgba(0,0,0,.3)}@media (prefers-color-scheme:dark){body{color:#fff;background:#000}.next-error-h1{border-right:1px solid rgba(255,255,255,.3)}}"}}],["$","h1",null,{"className":"next-error-h1","style":{"display":"inline-block","margin":"0 20px 0 0","padding":"0 23px 0 0","fontSize":24,"fontWeight":500,"verticalAlign":"top","lineHeight":"49px"},"children":"404"}],["$","div",null,{"style":{"display":"inline-block"},"children":["$","h2",null,{"style":{"fontSize":14,"fontWeight":400,"lineHeight":"49px","margin":0},"children":"This page could not be found."}]}]]}]}]],"notFoundStyles":[],"styles":null}]}]]}]]}],null]],[[["$","link","0",{"rel":"stylesheet","href":"/abe/group/_next/static/css/c1d59a1c2a22d6fb.css","precedence":"next","crossOrigin":""}]],"$L6"]]]] 7:I[1749,["749","static/chunks/749-849ff4ad7efada95.js","304","static/chunks/app/publications/page-3a6499dbd0295c42.js"],"Image"] 2:["$","div",null,{"className":"z-10 w-full max-w-4xl px-5 xl:px-0 flex flex-col items-center mt-16","children":[["$","div",null,{"className":"mb-7","children":[["$","h1",null,{"className":"animate-fade-up text-center font-display text-4xl font-bold tracking-[-0.04em] opacity-0 drop-shadow-sm [text-wrap:balance] md:text-7xl md:leading-[5rem]","style":{"animationDelay":"0.15s","animationFillMode":"forwards"},"children":"Publications"}],["$","div",null,{"className":"-mt-4","children":["$","div",null,{"className":"mt-6 animate-fade-up text-center text-gray-500 opacity-0 [text-wrap:balance] text-base","style":{"animationDelay":"0.25s","animationFillMode":"forwards"},"children":["$","p",null,{"children":"A (probably not up to date) list of our publications."}]}]}]]}],["$","div",null,{"className":"animate-fade-up opacity-0 flex flex-col space-y-10 relative max-w-3xl","style":{"animationDelay":"0.45s","animationFillMode":"forwards"},"children":[["$","div","eventfulness2023",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","eventfulness2023",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/eventfulness2023/./eventfulness.jpg","alt":"Cover Image for Eventfulness for Interactive Video Alignment","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"SIGGRAPH 2023 (Journal Paper)"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Eventfulness for Interactive Video Alignment"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Jiatian Sun, Longxiulin Deng, Triantafyllos Afouras, Andrew Owens and Abe Davis"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\nMore info/video/website/code coming soon!\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://dl.acm.org/doi/abs/10.1145/3592118","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://dl.acm.org/doi/abs/10.1145/3592118","children":"website"}]}],"$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","factormatte",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","factormatte",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/factormatte/./factormatte.jpg","alt":"Cover Image for FactorMatte: Redefining Video Matting for Re-Composition Tasks","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"SIGGRAPH 2023 (Journal Paper)"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"FactorMatte: Redefining Video Matting for Re-Composition Tasks"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Zeqi Gu, Wenqi Xian1, Noah Snavely, Abe Davis"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Video matting for scenes with complex interactions!\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://arxiv.org/abs/2211.02145","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://factormatte.github.io/","children":"website"}]}],"$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","iccv2023",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","iccv2023",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/iccv2023/./iccv2023.jpg","alt":"Cover Image for Ray Conditioning: Trading Photo-consistency for Photo-realism in Multi-view Image Generation","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"ICCV 2023"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Ray Conditioning: Trading Photo-consistency for Photo-realism in Multi-view Image Generation"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Eric Ming Chen, Sidhanth Holalkere, Ruyu Yan, Kai Zhang, Abe Davis"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"A nice clean light-weight trick for photorealistic viewpoint control over generative image models!\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://arxiv.org/abs/2304.13681","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://ray-cond.github.io/","children":"website"}]}],"$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","institchesCHI23",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","institchesCHI23",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/institchesCHI23/./institchesCHI23.jpg","alt":"Cover Image for InStitches: Augmenting Sewing Patterns with Personalized Material-Efficient Practice","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"CHI 2023"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"InStitches: Augmenting Sewing Patterns with Personalized Material-Efficient Practice"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Mackenzie Leake, Kathryn Jin, Abe Davis Stefanie Mueller."}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://hcie.csail.mit.edu/research/institches/institches.html","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://hcie.csail.mit.edu/research/institches/institches.html","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/NZmCreYYTJY","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","recapture",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","recapture",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/recapture/./recapture.jpg","alt":"Cover Image for ReCapture: AR-Guided Time-lapse Photography","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"UIST 2022"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"ReCapture: AR-Guided Time-lapse Photography"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Ruyu Yan, Jiatian Sun, Longxiulin Deng, Abe Davis"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\nAR-based guidance for hand-held time lapse capture.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://https://www.cs.cornell.edu/abe/projects/recapture/sources/ReCaptureUist2022.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://https://www.cs.cornell.edu/abe/projects/recapture/","children":"website"}]}],"","$undefined","$undefined"]}]]}]}]]}]}],["$","div","PaperPiecing2021",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","PaperPiecing2021",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/PaperPiecing2021/./PaperPiecing2021.jpg","alt":"Cover Image for A Mathematical Foundation for Foundation Paper Pieceable Quilts","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"SIGGRAPH 2021"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"A Mathematical Foundation for Foundation Paper Pieceable Quilts"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Mackenzie Leake, Gilbert Bernstein, Abe Davis, Maneesh Agrawala"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Computational tools for designing paper-pieceable quilts!\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":[["$","span",null,{"className":"bg-purple-50 text-purple-500 ring-purple-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://fpptool.herokuapp.com/","children":"demo"}]}],["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://web.stanford.edu/~mleake/projects/paperpiecing/files/FPP_small.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://web.stanford.edu/~mleake/projects/paperpiecing/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/g04VgzzRhlQ","children":"video"}]}],["$","span",null,{"className":"bg-red-50 text-red-700 ring-red-600/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://github.com/mleake/flipper","children":"code"}]}],"$undefined"]}]]}]}]]}]}],["$","div","crowdsamplenopt",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","crowdsamplenopt",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/crowdsamplenopt/./crowdsamplenopt.jpg","alt":"Cover Image for Crowdsampling the Plenoptic Function","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"ECCV 2020 (Selected for Oral Presentation)"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Crowdsampling the Plenoptic Function"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Zhengqi Li, Wenqi Xian, Abe Davis, and Noah Snavely"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Learn from Internet photos to create interactive space-time views of popular attractions! With self-supervision!\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://arxiv.org/abs/2007.15194","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://research.cs.cornell.edu/crowdplenoptic/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/MAVFKWX8LYo","children":"video"}]}],["$","span",null,{"className":"bg-red-50 text-red-700 ring-red-600/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://github.com/zhengqili/Crowdsampling-the-Plenoptic-Function","children":"code"}]}],"$undefined"]}]]}]}]]}]}],["$","div","vischirality",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","vischirality",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/vischirality/./vischirality.jpg","alt":"Cover Image for Visual Chirality","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"CVPR 2020 (Selected for Oral Presentation)"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Visual Chirality"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Zhiqiu Lin, Jin Sun, Abe Davis, Noah Snavely"}],["$","div",null,{"children":[["$","div",null,{"className":"inline-flex items-center rounded-lg px-2 py-1 font-bold bg-yellow-100 text-base text-yellow-700 ring-1 ring-inset ring-orange-600/10 text-sm","children":"Nominated for Best Paper"}]]}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\nHow does reflection change what we learn from images? Despite widespread use in data augmentation, people had not looked closely at this question before our work.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://arxiv.org/abs/2006.09512","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://https://linzhiqiu.github.io/papers/chirality/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/gc5IvTozU9M","children":"video"}]}],["$","span",null,{"className":"bg-red-50 text-red-700 ring-red-600/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://github.com/linzhiqiu/digital_chirality","children":"code"}]}],"$undefined"]}]]}]}]]}]}],["$","div","HMDCVPR2020Workshop",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","HMDCVPR2020Workshop",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/HMDCVPR2020Workshop/./HMDCVPR2020Workshop.jpg","alt":"Cover Image for Head-mounted Augmented Reality for Guided Surface Reflectance Capture","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"CVPR Workshop on Computer Vision for Augmented and Virtual Reality"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Head-mounted Augmented Reality for Guided Surface Reflectance Capture"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Harald Haraldsson, Søren Skovsen, Ser-Nam Lim, Steve Marschner, Serge Belongie, Abe Davis"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":""}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined","$undefined","$undefined",["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://https://youtu.be/dY7RRSDQc0Y","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","LocalAndSensingHMDCVPR2020XRWorkshop",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","LocalAndSensingHMDCVPR2020XRWorkshop",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/LocalAndSensingHMDCVPR2020XRWorkshop/./LocalAndSensingHMDCVPR2020XRWorkshop.jpg","alt":"Cover Image for Decoupled Localization and Sensing with HMD-based AR for Interactive Scene Acquisition","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"CVPR Workshop on Computer Vision for Augmented and Virtual Reality"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Decoupled Localization and Sensing with HMD-based AR for Interactive Scene Acquisition"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Søren Skovsen, Harald Haraldsson, Abe Davis, Henrik Karstoft, Serge Belongie"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":""}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined","$undefined","$undefined",["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://https://youtu.be/mFF7GWBsH00","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","iwshm2019",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","iwshm2019",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/iwshm2019/./iwshm2019.jpg","alt":"Cover Image for Modal Imaging of Portsmouth, New Hampshire Bridge for Video-Based Structural Health Monitoring","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"IWSHM 2019"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Modal Imaging of Portsmouth, New Hampshire Bridge for Video-Based Structural Health Monitoring"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis, Justin G. Chen, Abe Davis, Travis M. Adams, Hao Sun, Erin S. Bell, Fredo Durand, Oral Buyukozturk"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Comparing modal images of a bridge with an FEM model of that bridge.\n\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/files/papers/DavisIWSHM19.pdf","children":"paper"}]}],"$undefined","$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","visbeat",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","visbeat",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/visbeat/./visbeat.jpg","alt":"Cover Image for Visual Rhythm and Beat","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"SIGGRAPH 2018"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Visual Rhythm and Beat"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis and Maneesh Agrawala"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Make arbitrary video dance to arbitrary music.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://www.abedavis.com/files/papers/VisualRhythm_Davis18.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/visualbeat/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/K3z68mOLbNo","children":"video"}]}],["$","span",null,{"className":"bg-red-50 text-red-700 ring-red-600/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://github.com/abedavis/visbeat","children":"code"}]}],"$undefined"]}]]}]}]]}]}],["$","div","iwshm2017",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","iwshm2017",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/iwshm2017/./iwshm2017.jpg","alt":"Cover Image for Structural Health Monitoring from the Window Seat of a Passenger Airplane","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"IWSHM 2017"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Structural Health Monitoring from the Window Seat of a Passenger Airplane"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis, Justin G. Chen, Oral Buyukozturk, Fredo Durand, Doug L. James"}],["$","div",null,{"children":[["$","div",null,{"className":"inline-flex items-center rounded-lg px-2 py-1 font-bold bg-yellow-100 text-base text-yellow-700 ring-1 ring-inset ring-orange-600/10 text-sm","children":"Winner, SHM In Action 2017"}]]}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Monitoring the vibration modes of an airplane wing from the window seat using a mobile phone.\n\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/files/papers/AbeDavisIWSHM2017.pdf","children":"paper"}]}],"$undefined","$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","roughcut",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","roughcut",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/roughcut/./roughcut.jpg","alt":"Cover Image for Computational Video Editing for Dialogue-Driven Scenes","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"SIGGRAPH 2017"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Computational Video Editing for Dialogue-Driven Scenes"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Mackenzie Leake, Abe Davis, Anh Truong and Maneesh Agrawala"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Computational tools for fast, high-level, editing of multiple video takes for dialogue driven scenes.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://graphics.stanford.edu/papers/roughcut/files/roughcut-small.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://graphics.stanford.edu/papers/roughcut/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/tF43Zqoue20","children":"video"}]}],"$undefined",["$","span",null,{"className":"bg-pink-50 text-pink-700 ring-pink-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://graphics.stanford.edu/papers/roughcut/supplementary/index.html","children":"supplemental"}]}]]}]]}]}]]}]}],["$","div","bounceflash",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","bounceflash",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/bounceflash/./bounceflash.jpg","alt":"Cover Image for Computational Bounce Flash for Indoor Portraits","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"SIGGRAPH Asia 2016"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Computational Bounce Flash for Indoor Portraits"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Lukas Murmann, Abe Davis, Jan Kautz, Fredo Durand"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\nA robotic bounce flash that makes portraiture lighting super easy\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/files/papers/BounceFlash_SIGA16.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://lmurmann.net/bounceflash","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://https://www.youtube.com/watch?v=Au44yDS6yZE","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","visvib_pami",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","visvib_pami",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/visvib_pami/./visvib_pami.jpg","alt":"Cover Image for Visual Vibrometry: Estimating Material Properties from Small Motions in Video","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"PAMI 2016"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Visual Vibrometry: Estimating Material Properties from Small Motions in Video"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis*, Katherine L. Bouman*, Justin G. Chen, Michael Rubinstein, Oral Buyukozturk, Fredo Durand, William T. Freeman"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Improving on our CVPR 2015 paper, we adding additional analysis, showed you could do damping estimation, and infer hidden structure (e.g., tell whether a container is leaking).\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/files/papers/visvib_pami.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://www.visualvibrometry.com","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://https://youtu.be/5apFqYEx5ew","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","wcndt16",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","wcndt16",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/wcndt16/./wcndt16.jpg","alt":"Cover Image for Smaller Than the Eye Can See: Vibration Analysis with Video Cameras","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"19th World Conference on Non-Destructive Testing 2016 (WCNDT)"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Smaller Than the Eye Can See: Vibration Analysis with Video Cameras"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Oral Buyukozturk, Justin G. Chen, Neal Wadhwa, Abe Davis, Frédo Durand, and William T. Freeman"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Including visual vibration analysis of earthquakes!\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://http://people.csail.mit.edu/ju21743/docs/buyukozturk_WCNDT_2016.pdf","children":"paper"}]}],"$undefined","$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","phdthesis",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","phdthesis",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/phdthesis/./phdthesis.jpg","alt":"Cover Image for Visual Vibration Analysis","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"MIT EECS PhD Thesis"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Visual Vibration Analysis"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis"}],["$","div",null,{"children":[["$","div",null,{"className":"inline-flex items-center rounded-lg px-2 py-1 font-bold bg-yellow-100 text-base text-yellow-700 ring-1 ring-inset ring-orange-600/10 text-sm","children":"Winner of the MIT Sprowls thesis award"}],["$","div",null,{"className":"inline-flex items-center rounded-lg px-2 py-1 font-bold bg-yellow-100 text-base text-yellow-700 ring-1 ring-inset ring-orange-600/10 text-sm","children":"ACM SIGGRAPH Thesis Award, Honorable mention"}]]}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"My dissertation on visual vibration analysis and its applications.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/files/papers/thesis.pdf","children":"paper"}]}],"$undefined","$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","ismb",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","ismb",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/ismb/./ismb.jpg","alt":"Cover Image for \"Image-Space Modal Bases for Plausible Manipulation of Objects in Video\"","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"SIGGRAPH Asia 2015"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"\"Image-Space Modal Bases for Plausible Manipulation of Objects in Video\""}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis, Justin Chen, Fredo Durand"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\nThis paper describes an improved version of the demo I showed at the end of my TED 2015 talk (with some improvements since TED) and demonstrates applications related to low-cost special effects.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/files/papers/ISMB_Davis_2015.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://www.interactivedynamicvideo.com/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/4f09VdXex3A","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","isndtce_1",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","isndtce_1",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/isndtce_1/./isndtce_1.jpg","alt":"Cover Image for Video Camera-based Vibration Measurement for Condition Assessment of Civil Infrastructure","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"International Symposium Non-Destructive Testing in Civil Engineering"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Video Camera-based Vibration Measurement for Condition Assessment of Civil Infrastructure"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Chen, J. G.; Davis, A.; Wadhwa, N.; Durand, F.; Freeman, W. T.; Buyukozturk, O."}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\nMore work with folks from the Civil Engineering department on NDT of structures using visual vibration analysis.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://people.csail.mit.edu/ju21743/docs/chen_NDTCE_2015.pdf","children":"paper"}]}],"$undefined","$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","iwshm15",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","iwshm15",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/iwshm15/./iwshm15.jpg","alt":"Cover Image for Long Distance Video Camera Measurements Of Structures","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"10th International Workshop on Structural Health Monitoring (IWSHM 2015)"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Long Distance Video Camera Measurements Of Structures"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Justin G. Chen, Neal Wadhwa, Abe Davis, Fredo Durand, William T. Freeman, Oral Buyukozturk"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Work with folks from the Civil Engineering department looking at how distance effects visual vibration analysis for NDT of structures. I presented this at IWSHM 2015.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://people.csail.mit.edu/ju21743/docs/chen_IWSHM_2015.pdf","children":"paper"}]}],"$undefined","$undefined","$undefined","$undefined"]}]]}]}]]}]}],["$","div","sparselightfields",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","sparselightfields",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/sparselightfields/./sparselightfields.jpg","alt":"Cover Image for Light Field Reconstruction Using Sparsity in the Continuous Fourier Domain","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"TOG / SIGGRAPH 2015"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Light Field Reconstruction Using Sparsity in the Continuous Fourier Domain"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Lixin Shi, Haitham Hassanieh, Abe Davis, Dina Katabi, Fredo Durand"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\n1) We show that light fields are really only sparse in the continuous frequency domain,\nand 2) We show how to leverage that sparsity in an optimization over continuous frequencies to reconstruct even highly non-Lambertian light fields.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://groups.csail.mit.edu/netmit/LFSparseRecon/files/a12-shi.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://groups.csail.mit.edu/netmit/LFSparseRecon/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/SPKEzr5Lh1g","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","visvib_cvpr",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","visvib_cvpr",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/visvib_cvpr/./visvib_cvpr.jpg","alt":"Cover Image for Visual Vibrometry: Estimating Material Properties from Small Motions in Video","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"CVPR 2015 (Selected for Oral Presentation)"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Visual Vibrometry: Estimating Material Properties from Small Motions in Video"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis*, Katherine L. Bouman*, Justin G. Chen, Michael Rubinstein, Fredo Durand, William T. Freeman"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/files/papers/visvib_cvpr15.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://www.visualvibrometry.com/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://https://youtu.be/5apFqYEx5ew","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","vismic",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","vismic",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/vismic/./vismic.jpg","alt":"Cover Image for The Visual Microphone: Passive Recovery of Sound from Video","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"SIGGRAPH 2014"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"The Visual Microphone: Passive Recovery of Sound from Video"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis, Michael Rubinstein, Neal Wadhwa, Gautham J. Mysore, Fredo Durand, William T. Freeman"}],["$","div",null,{"children":[["$","div",null,{"className":"inline-flex items-center rounded-lg px-2 py-1 font-bold bg-yellow-100 text-base text-yellow-700 ring-1 ring-inset ring-orange-600/10 text-sm","children":"Discover Magazine Best Science Breakthroughs of 2014, #32"}]]}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"This project received a lot of press. Check out the project webpage for some examples. Also check out the Code.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://people.csail.mit.edu/mrub/papers/VisualMic_SIGGRAPH2014.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://people.csail.mit.edu/mrub/VisualMic/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtu.be/FKXOucXB4a8","children":"video"}]}],["$","span",null,{"className":"bg-red-50 text-red-700 ring-red-600/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://people.csail.mit.edu/abedavis/research/VisMic/VMSlim.zip","children":"code"}]}],"$undefined"]}]]}]}]]}]}],["$","div","caperture",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","caperture",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/caperture/./caperture.jpg","alt":"Cover Image for Caperture","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"iOS App"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Caperture"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"I built on my masters thesis to create an AR-based iOS app for capturing light fields. Caperture no longer exists, but it was one of the first AR apps on a mobile phone (years before apple introduced ARKit).\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined","$undefined",["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://abedavis.com/caperture/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtube.com/watch?v=hRIgup3KYMA","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","speckle_cvpr12",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","speckle_cvpr12",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/speckle_cvpr12/./speckle_cvpr12.jpg","alt":"Cover Image for Laser Speckle Photography for Surface Tampering Detection","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"IEEE Conference on Computer Vision and Pattern Recognition, (CVPR 2012)"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Laser Speckle Photography for Surface Tampering Detection"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"YiChang Shih, Abe Davis, Samuel W. Hasinoff, Fredo Durand, and William T. Freeman"}],["$","div",null,{"children":[["$","div",null,{"className":"inline-flex items-center rounded-lg px-2 py-1 font-bold bg-yellow-100 text-base text-yellow-700 ring-1 ring-inset ring-orange-600/10 text-sm","children":"Google Student Travel Award"}]]}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"Use lasers to detect tampering with a surface. We can detect whether someone has lightly touched a surface even if they were wearing gloves! This system also uses AR for ultra-precise rephotography.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://groups.csail.mit.edu/graphics/speckle/speckle_final_low_res.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://groups.csail.mit.edu/graphics/speckle/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtube.com/watch?v=tYFLze9VwB0","children":"video"}]}],"$undefined",["$","span",null,{"className":"bg-pink-50 text-pink-700 ring-pink-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://groups.csail.mit.edu/graphics/speckle/supp.pdf","children":"supplemental"}]}]]}]]}]}]]}]}],["$","div","ulf",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","ulf",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/ulf/./ulf.jpg","alt":"Cover Image for Unstructured Light Fields","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"Eurographics 2012"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Unstructured Light Fields"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Abe Davis, Marc Levoy, Fredo Durand"}],["$","div",null,{"children":"$undefined"}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\nI turned this into a free iOS app called Caperture.\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://people.csail.mit.edu/abedavis/ULF/DavisEG2012.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://people.csail.mit.edu/abedavis/ULF/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtube.com/watch?v=WU-D4PA5plo","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}],["$","div","permuto_eg10",{"className":"flex flex-col items-center transition-all hover:scale-110","children":["$","div",null,{"className":"flex soft-shadow rounded-2xl w-full items-stretch overflow-hidden bg-white","children":[["$","div","permuto_eg10",{"className":"overflow-hidden min-h-full flex flex-none relative w-1/3","children":["$","$L7",null,{"src":"/abe/group/content/publications/permuto_eg10/./permuto_eg10.png","alt":"Cover Image for Fast High-Dimensional Filtering Using the Permutohedral Lattice","fill":true,"style":{"objectFit":"contain","padding":"10px"},"priority":true}]}],["$","div",null,{"children":["$","div",null,{"className":"flex flex-col space-y-3 justify-between h-full p-6","children":[["$","div",null,{"className":"flex flex-col space-y-0.5","children":[["$","div",null,{"children":[["$","time",null,{"className":"inline-flex text-xs text-gray-500"}]," ",["$","span",null,{"className":"inline-flex text-xs text-gray-600","children":"Eurographics 2010"}]]}],["$","h3",null,{"className":"text-lg font-bold","children":"Fast High-Dimensional Filtering Using the Permutohedral Lattice"}],["$","h3",null,{"className":"text-xs text-gray-700","children":"Andrew Adams, Jongmin Baek, Abe Davis (as Myers Abraham Davis)"}],["$","div",null,{"children":[["$","div",null,{"className":"inline-flex items-center rounded-lg px-2 py-1 font-bold bg-yellow-100 text-base text-yellow-700 ring-1 ring-inset ring-orange-600/10 text-sm","children":"Runner-up, Best Paper Award"}]]}],["$","div",null,{"className":"publication-card-markdown text-xs text-gray-400 leading-relaxed","dangerouslySetInnerHTML":{"__html":"\n"}}]]}],["$","div",null,{"className":"flex space-x-2","children":["$undefined",["$","span",null,{"className":"bg-seagull-100 text-seagull-600 ring-seagull-700/10 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://graphics.stanford.edu/papers/permutohedral/permutohedral.pdf","children":"paper"}]}],["$","span",null,{"className":"bg-green-50 text-green-700 ring-green-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://graphics.stanford.edu/papers/permutohedral/","children":"website"}]}],["$","span",null,{"className":"bg-yellow-50 text-yellow-800 ring-yellow-600/20 inline-flex items-center rounded-md text-xs font-medium ring-1 ring-inset px-2 py-1","children":["$","a",null,{"href":"https://youtube.com/watch?v=e7kLRllwHPc","children":"video"}]}],"$undefined","$undefined"]}]]}]}]]}]}]]}]]}] 6:[["$","meta","0",{"name":"viewport","content":"width=device-width, initial-scale=1"}],["$","meta","1",{"charSet":"utf-8"}],["$","title","2",{"children":"Abe Davis's Group"}],["$","meta","3",{"name":"description","content":"Detailed information about Abe Davis's Research Group."}],["$","meta","4",{"name":"next-size-adjust"}]] 1:null