1:"$Sreact.fragment"
2:I[5244,[],""]
3:I[3866,[],""]
4:I[8173,["173","static/chunks/173-38e78c9463adde84.js","736","static/chunks/app/resume/page-1e711bec227bed46.js"],""]
5:I[6213,[],"OutletBoundary"]
7:I[6213,[],"MetadataBoundary"]
9:I[6213,[],"ViewportBoundary"]
b:I[4835,[],""]
:HL["/~bryanw/_next/static/media/30d74baa196fe88a-s.p.woff2","font",{"crossOrigin":"","type":"font/woff2"}]
:HL["/~bryanw/_next/static/media/4cf2300e9c8272f7-s.p.woff2","font",{"crossOrigin":"","type":"font/woff2"}]
:HL["/~bryanw/_next/static/media/886f446b96dc7734-s.p.woff2","font",{"crossOrigin":"","type":"font/woff2"}]
:HL["/~bryanw/_next/static/media/93f479601ee12b01-s.p.woff2","font",{"crossOrigin":"","type":"font/woff2"}]
:HL["/~bryanw/_next/static/media/e693e841d50dcf2f-s.p.woff2","font",{"crossOrigin":"","type":"font/woff2"}]
:HL["/~bryanw/_next/static/css/aa8f61e79074ca38.css","style"]
:HL["/~bryanw/_next/static/css/57827a59dd5eb61a.css","style"]
0:{"P":null,"b":"lPdEByy3309p_SN6-tpcS","p":"/~bryanw","c":["","resume"],"i":false,"f":[[["",{"children":["resume",{"children":["__PAGE__",{}]}]},"$undefined","$undefined",true],["",["$","$1","c",{"children":[[["$","link","0",{"rel":"stylesheet","href":"/~bryanw/_next/static/css/aa8f61e79074ca38.css","precedence":"next","crossOrigin":"$undefined","nonce":"$undefined"}]],["$","html",null,{"lang":"en","children":["$","body",null,{"className":"__variable_c6514f __variable_5a4633 __variable_1d793e __variable_08fb03 antialiased","children":[["$","main",null,{"className":"","children":["$","$L2",null,{"parallelRouterKey":"children","segmentPath":["children"],"error":"$undefined","errorStyles":"$undefined","errorScripts":"$undefined","template":["$","$L3",null,{}],"templateStyles":"$undefined","templateScripts":"$undefined","notFound":[[],[["$","title",null,{"children":"404: This page could not be found."}],["$","div",null,{"style":{"fontFamily":"system-ui,\"Segoe UI\",Roboto,Helvetica,Arial,sans-serif,\"Apple Color Emoji\",\"Segoe UI Emoji\"","height":"100vh","textAlign":"center","display":"flex","flexDirection":"column","alignItems":"center","justifyContent":"center"},"children":["$","div",null,{"children":[["$","style",null,{"dangerouslySetInnerHTML":{"__html":"body{color:#000;background:#fff;margin:0}.next-error-h1{border-right:1px solid rgba(0,0,0,.3)}@media (prefers-color-scheme:dark){body{color:#fff;background:#000}.next-error-h1{border-right:1px solid rgba(255,255,255,.3)}}"}}],["$","h1",null,{"className":"next-error-h1","style":{"display":"inline-block","margin":"0 20px 0 0","padding":"0 23px 0 0","fontSize":24,"fontWeight":500,"verticalAlign":"top","lineHeight":"49px"},"children":404}],["$","div",null,{"style":{"display":"inline-block"},"children":["$","h2",null,{"style":{"fontSize":14,"fontWeight":400,"lineHeight":"49px","margin":0},"children":"This page could not be found."}]}]]}]}]]],"forbidden":"$undefined","unauthorized":"$undefined"}]}],["$","footer",null,{"className":"border-t border-neutral-200 dark:border-neutral-800 bg-[#FFFCF8]","children":["$","div",null,{"className":"flex flex-row mx-auto max-w-7xl px-6 py-12 md:flex md:items-start md:justify-between ","children":[["$","div",null,{"className":"mb-4 text-sm text-neutral-600 dark:text-neutral-400","children":[["$","p",null,{"children":["© ",2026," ","Bryan Wang, Ph.D.","."]}],"$undefined"]}],["$","div",null,{"className":"mb-4","children":["$","p",null,{"className":"text-sm text-neutral-500 dark:text-neutral-500 justify","children":["Built with"," ",["$","a",null,{"href":"https://github.com/tovacinni/research-website-template","className":"underline hover:text-neutral-800 dark:hover:text-neutral-300 transition-colors","children":"research-website-template"}]]}]}]]}]}]]}]}]]}],{"children":["resume",["$","$1","c",{"children":[null,["$","$L2",null,{"parallelRouterKey":"children","segmentPath":["children","resume","children"],"error":"$undefined","errorStyles":"$undefined","errorScripts":"$undefined","template":["$","$L3",null,{}],"templateStyles":"$undefined","templateScripts":"$undefined","notFound":"$undefined","forbidden":"$undefined","unauthorized":"$undefined"}]]}],{"children":["__PAGE__",["$","$1","c",{"children":[["$","div",null,{"className":"min-h-screen bg-[#FFFCF8] resume-page","children":[["$","style",null,{"children":"@media print { @page { size: A4; margin: 0.5in 0.45in; } }"}],["$","div",null,{"className":"max-w-3xl mx-auto px-6 md:px-10 py-10 md:py-14 print-doc","children":[["$","div",null,{"className":"print:hidden flex items-center justify-between mb-10 pb-4 border-b border-zinc-200","children":[["$","$L4",null,{"href":"/","className":"inline-flex items-center gap-2 text-xs tracking-wider uppercase text-zinc-500 hover:text-zinc-900 transition-colors","children":[["$","svg",null,{"ref":"$undefined","xmlns":"http://www.w3.org/2000/svg","width":12,"height":12,"viewBox":"0 0 24 24","fill":"none","stroke":"currentColor","strokeWidth":2,"strokeLinecap":"round","strokeLinejoin":"round","className":"lucide lucide-arrow-left","children":[["$","path","1l729n",{"d":"m12 19-7-7 7-7"}],["$","path","x3x0zl",{"d":"M19 12H5"}],"$undefined"]}]," Back to site"]}],["$","a",null,{"href":"/~bryanw/bryan-wang-resume.pdf","download":true,"className":"inline-flex items-center gap-2 px-3 py-1.5 text-xs font-medium tracking-wider uppercase border border-zinc-300 rounded-full text-zinc-700 hover:bg-zinc-900 hover:text-white hover:border-zinc-900 transition-colors duration-200","aria-label":"Download résumé as PDF","children":[["$","svg",null,{"ref":"$undefined","xmlns":"http://www.w3.org/2000/svg","width":12,"height":12,"viewBox":"0 0 24 24","fill":"none","stroke":"currentColor","strokeWidth":2,"strokeLinecap":"round","strokeLinejoin":"round","className":"lucide lucide-download","children":[["$","path","ih7n3h",{"d":"M21 15v4a2 2 0 0 1-2 2H5a2 2 0 0 1-2-2v-4"}],["$","polyline","2ggqvy",{"points":"7 10 12 15 17 10"}],["$","line","1vk2je",{"x1":"12","x2":"12","y1":"15","y2":"3"}],"$undefined"]}],"Download PDF"]}]]}],["$","header",null,{"className":"mb-6 avoid-break","children":[["$","h1",null,{"className":"font-serif text-3xl tracking-wide text-zinc-900 mb-1","children":"Bryan Wang"}],["$","p",null,{"className":"text-sm text-zinc-700 mb-1","children":["Research Scientist",[" · ","Adobe Research"]]}],["$","p",null,{"className":"text-xs text-zinc-500 mb-2","children":"Ph.D. in Computer Science · University of Toronto"}],["$","div",null,{"className":"flex flex-wrap gap-x-4 gap-y-1 text-xs text-zinc-600","children":[["$","a",null,{"className":"print-link","href":"mailto:b02902096@gmail.com","children":"b02902096@gmail.com"}],["$","a",null,{"className":"print-link","href":"https://www.dgp.toronto.edu/~bryanw/","target":"_blank","rel":"noopener noreferrer","children":"Website"}],["$","a",null,{"className":"print-link","href":"https://scholar.google.com/citations?user=2s6wkyYAAAAJ&hl","target":"_blank","rel":"noopener noreferrer","children":"Google Scholar"}],["$","a",null,{"className":"print-link","href":"https://www.linkedin.com/in/bryan-wang-a97276119","target":"_blank","rel":"noopener noreferrer","children":"LinkedIn"}],"$undefined",["$","a",null,{"className":"print-link","href":"https://twitter.com/bryanhaoenwang","target":"_blank","rel":"noopener noreferrer","children":"Twitter"}]]}]]}],["$","section",null,{"className":"mb-6","children":[["$","h2",null,{"className":"font-serif text-sm tracking-[0.2em] uppercase text-zinc-800 border-b border-zinc-300 pb-1 mb-3","children":"Summary"}],["$","div",null,{"className":"space-y-2","children":[["$","p","0",{"className":"text-sm text-zinc-700 leading-relaxed","children":"Human-AI interaction researcher who ships AI-powered tools that augment human creativity and capability, with a recent focus on generative AI for video and audio. End-to-end builder across ML, backend, frontend, and interaction design."}]]}]]}],["$","section",null,{"className":"mb-6","children":[["$","h2",null,{"className":"font-serif text-sm tracking-[0.2em] uppercase text-zinc-800 border-b border-zinc-300 pb-1 mb-3","children":"Experience"}],["$","ul",null,{"className":"space-y-3","children":[["$","li",null,{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900","children":["Research Scientist",[["$","span",null,{"className":"text-zinc-400","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"Adobe Research"}]]]}],["$","p",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2024 to Present"}]]}],["$","p",null,{"className":"text-sm text-zinc-700 leading-relaxed mt-1","children":"Leading 0→1 research on AI creative tools, from problem framing to prototyping and shipping, in close collaboration with Product, Design, and Engineering."}],["$","ul",null,{"className":"mt-1.5 space-y-1 text-sm text-zinc-700 leading-relaxed list-disc pl-5 marker:text-zinc-400","children":[["$","li",null,{"children":"Led Precision Flow and contributed to Generate Soundtrack, both shipped in Adobe Firefly"}],["$","li",null,{"children":"Mentored 4 PhD interns, published 6 peer-reviewed papers, and filed 3 patents"}]]}]]}],[["$","li","0",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900","children":["Research Intern",[["$","span",null,{"className":"text-zinc-400","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"Meta Reality Labs - Research"}]]]}],["$","p",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"Summer 2023"}]]}],["$","p",null,{"className":"text-sm text-zinc-700 leading-relaxed mt-1","children":"Created LAVE, an LLM-powered video editing agent that plans, searches, and edits footage through natural conversation."}]]}],["$","li","1",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900","children":["Student Researcher",[["$","span",null,{"className":"text-zinc-400","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"Google Research"}]]]}],["$","p",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"Summer 2022"}]]}],["$","p",null,{"className":"text-sm text-zinc-700 leading-relaxed mt-1","children":"Among the first to show LLMs could enable conversational interaction with GUI through prompting alone, no task-specific training required."}]]}],["$","li","2",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900","children":["Research Intern",[["$","span",null,{"className":"text-zinc-400","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"Adobe Research"}]]]}],["$","p",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"Summer 2021, Winter 2022"}]]}],["$","p",null,{"className":"text-sm text-zinc-700 leading-relaxed mt-1","children":"Built ROPE, an NLP-driven tool that automatically condenses voice recordings into social-media-ready audio stories while preserving narrative flow."}]]}],["$","li","3",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900","children":["Research Intern",[["$","span",null,{"className":"text-zinc-400","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"Google Research"}]]]}],["$","p",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"Summer 2020, Winter 2021"}]]}],["$","p",null,{"className":"text-sm text-zinc-700 leading-relaxed mt-1","children":"Created a 110K-summary, 22K-screen multimodal dataset and trained vision-language models to automatically summarize mobile UIs."}]]}]]]}]]}],["$","section",null,{"className":"mb-6","children":[["$","h2",null,{"className":"font-serif text-sm tracking-[0.2em] uppercase text-zinc-800 border-b border-zinc-300 pb-1 mb-3","children":"Shipped Products"}],["$","ul",null,{"className":"space-y-3","children":[["$","li","0",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900","children":[["$","a",null,{"className":"print-link text-zinc-900","href":"https://blog.adobe.com/en/publish/2026/04/09/new-image-editing-features-adobe-firefly-get-you-from-almost-there-to-exactly-right","target":"_blank","rel":"noopener noreferrer","children":"Precision Flow"}],["$","span",null,{"className":"text-zinc-400","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"Adobe Firefly"}],["$","span",null,{"className":"text-xs text-zinc-500","children":[" · ","Research Lead (0→1)"]}]]}],["$","p",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2026"}]]}],["$","p",null,{"className":"text-sm text-zinc-700 leading-relaxed mt-1","children":"A Firefly image editing feature that renders every edit as a continuous spectrum. Glide a slider from A to B and land anywhere in between."}]]}],["$","li","1",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900","children":[["$","a",null,{"className":"print-link text-zinc-900","href":"https://www.theverge.com/news/807809/adobe-firefly-ai-audio-generate-soundtrack-speech","target":"_blank","rel":"noopener noreferrer","children":"Generate Soundtrack"}],["$","span",null,{"className":"text-zinc-400","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"Adobe Firefly"}],["$","span",null,{"className":"text-xs text-zinc-500","children":[" · ","Contributor"]}]]}],["$","p",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2025"}]]}],["$","p",null,{"className":"text-sm text-zinc-700 leading-relaxed mt-1","children":"A Firefly audio tool that generates commercially-safe instrumental soundtracks from a text prompt or video. Music that fits your footage, automatically."}]]}]]}]]}],["$","section",null,{"className":"mb-6","children":[["$","h2",null,{"className":"font-serif text-sm tracking-[0.2em] uppercase text-zinc-800 border-b border-zinc-300 pb-1 mb-3","children":"Selected Publications"}],[["$","p",null,{"className":"text-xs text-zinc-500 mb-3","children":["1,400+ citations · h-index 16 · See full list on"," ",["$","a",null,{"className":"print-link","href":"https://scholar.google.com/citations?user=2s6wkyYAAAAJ&hl","target":"_blank","rel":"noopener noreferrer","children":"Google Scholar"}],"."]}],["$","ul",null,{"className":"space-y-2.5","children":[["$","li","0",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900 leading-snug flex-1","children":["$","a",null,{"className":"print-link text-zinc-900","href":"https://arxiv.org/abs/2601.22013","target":"_blank","rel":"noopener noreferrer","children":"Vidmento: Creating Video Stories Through Context-Aware Expansion With Generative Video"}]}],["$","p",null,{"className":"text-xs text-zinc-500 italic tabular-nums whitespace-nowrap","children":["CHI"," ","2026"]}]]}],["$","p",null,{"className":"text-xs text-zinc-600 leading-snug mt-0.5","dangerouslySetInnerHTML":{"__html":"Catherine Yeh, Anh Truong, Mira Dontcheva, Bryan Wang"}}]]}],["$","li","1",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900 leading-snug flex-1","children":["$","a",null,{"className":"print-link text-zinc-900","href":"https://arxiv.org/abs/2601.12180","target":"_blank","rel":"noopener noreferrer","children":"VidTune: Creating Video Soundtracks with Generative Music and Contextual Thumbnails"}]}],["$","p",null,{"className":"text-xs text-zinc-500 italic tabular-nums whitespace-nowrap","children":["CHI"," ","2026"]}]]}],["$","p",null,{"className":"text-xs text-zinc-600 leading-snug mt-0.5","dangerouslySetInnerHTML":{"__html":"Mina Huh, Ailie C. Fraser, Dingzeyu Li, Mira Dontcheva, Bryan Wang"}}]]}],["$","li","2",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900 leading-snug flex-1","children":["$","a",null,{"className":"print-link text-zinc-900","href":"https://arxiv.org/abs/2505.21966","target":"_blank","rel":"noopener noreferrer","children":"MapStory: Prototyping Editable Map Animations with LLM Agents"}]}],["$","p",null,{"className":"text-xs text-zinc-500 italic tabular-nums whitespace-nowrap","children":["UIST"," ","2025"]}]]}],["$","p",null,{"className":"text-xs text-zinc-600 leading-snug mt-0.5","dangerouslySetInnerHTML":{"__html":"Aditya Gunturu, Ben Pearman, Keiichi Ihara, Morteza Faraji, Bryan Wang, Rubaiat Habib Kazi, Ryo Suzuki"}}]]}],["$","li","3",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900 leading-snug flex-1","children":["$","a",null,{"className":"print-link text-zinc-900","href":"https://arxiv.org/pdf/2402.10294","target":"_blank","rel":"noopener noreferrer","children":"LAVE: LLM-Powered Agent Assistance and Language Augmentation for Video Editing"}]}],["$","p",null,{"className":"text-xs text-zinc-500 italic tabular-nums whitespace-nowrap","children":["IUI"," ","2024"]}]]}],["$","p",null,{"className":"text-xs text-zinc-600 leading-snug mt-0.5","dangerouslySetInnerHTML":{"__html":"Bryan Wang, Yuliang Li, Zhaoyang Lv, Haijun Xia, Yan Xu, Raj Sodhi"}}]]}],["$","li","4",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900 leading-snug flex-1","children":["$","a",null,{"className":"print-link text-zinc-900","href":"/~bryanw/pdf/promptify.pdf","target":"_blank","rel":"noopener noreferrer","children":"Promptify: Text-to-Image Generation through Interactive Prompt Exploration with Large Language Models"}]}],["$","p",null,{"className":"text-xs text-zinc-500 italic tabular-nums whitespace-nowrap","children":["UIST"," ","2023"]}]]}],["$","p",null,{"className":"text-xs text-zinc-600 leading-snug mt-0.5","dangerouslySetInnerHTML":{"__html":"Stephen Brade, Bryan Wang, Mauricio Sousa, Sageev Oore, Tovi Grossman"}}]]}],["$","li","5",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900 leading-snug flex-1","children":["$","a",null,{"className":"print-link text-zinc-900","href":"/~bryanw/pdf/llm4ui.pdf","target":"_blank","rel":"noopener noreferrer","children":"Enabling Conversational Interaction with Mobile UI using Large Language Models"}]}],["$","p",null,{"className":"text-xs text-zinc-500 italic tabular-nums whitespace-nowrap","children":["CHI"," ","2023"]}]]}],["$","p",null,{"className":"text-xs text-zinc-600 leading-snug mt-0.5","dangerouslySetInnerHTML":{"__html":"Bryan Wang, Gang Li, Yang Li"}}]]}],["$","li","6",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","p",null,{"className":"text-sm font-medium text-zinc-900 leading-snug flex-1","children":["$","a",null,{"className":"print-link text-zinc-900","href":"/~bryanw/pdf/soloist.pdf","target":"_blank","rel":"noopener noreferrer","children":"Soloist: Generating Mixed-Initiative Tutorials from Existing Guitar Instructional Videos Through Audio Processing"}]}],["$","p",null,{"className":"text-xs text-zinc-500 italic tabular-nums whitespace-nowrap","children":["CHI"," ","2021"]}]]}],["$","p",null,{"className":"text-xs text-zinc-600 leading-snug mt-0.5","dangerouslySetInnerHTML":{"__html":"Bryan Wang, Meng Yu Yang, Tovi Grossman"}}]]}]]}]]]}],["$","section",null,{"className":"mb-6","children":[["$","h2",null,{"className":"font-serif text-sm tracking-[0.2em] uppercase text-zinc-800 border-b border-zinc-300 pb-1 mb-3","children":"Education"}],["$","ul",null,{"className":"space-y-2","children":[["$","li","0",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","span",null,{"className":"text-sm text-zinc-800","children":[["$","span",null,{"className":"font-medium","children":"Ph.D. in Computer Science"}],["$","span",null,{"className":"text-zinc-500","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"University of Toronto"}],["$","span",null,{"className":"text-xs text-zinc-500","children":[" · ","Advisor: ","Tovi Grossman"]}]]}],["$","span",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2024"}]]}],["$","p",null,{"className":"text-xs text-zinc-600 italic mt-0.5","children":["Thesis:"," ",["$","a",null,{"className":"print-link","href":"https://www.proquest.com/openview/cb5a45799cabb98e9297ce0756131258/1?pq-origsite=gscholar&cbl=18750&diss=y","target":"_blank","rel":"noopener noreferrer","children":"Human-AI Systems for Creating, Consuming, and Interacting with Digital Content"}]]}]]}],["$","li","1",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","span",null,{"className":"text-sm text-zinc-800","children":[["$","span",null,{"className":"font-medium","children":"M.Sc. in Computer Science"}],["$","span",null,{"className":"text-zinc-500","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"University of Toronto"}],["$","span",null,{"className":"text-xs text-zinc-500","children":[" · ","Advisor: ","Tovi Grossman"]}]]}],["$","span",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2020"}]]}],"$undefined"]}],["$","li","2",{"className":"avoid-break","children":[["$","div",null,{"className":"flex items-baseline justify-between gap-4","children":[["$","span",null,{"className":"text-sm text-zinc-800","children":[["$","span",null,{"className":"font-medium","children":"B.Sc. in Computer Science"}],["$","span",null,{"className":"text-zinc-500","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"National Taiwan University"}],"$undefined"]}],["$","span",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2018"}]]}],"$undefined"]}]]}]]}],["$","section",null,{"className":"mb-6","children":[["$","h2",null,{"className":"font-serif text-sm tracking-[0.2em] uppercase text-zinc-800 border-b border-zinc-300 pb-1 mb-3","children":"Selected Honours"}],["$","ul",null,{"className":"space-y-1.5","children":[["$","li","0",{"className":"avoid-break flex items-baseline justify-between gap-3 text-sm","children":[["$","span",null,{"className":"text-zinc-800","children":[["$","span",null,{"className":"font-medium","children":"Jazzki Human-Machine Creativity Award, First Place"}],"$undefined",["$","span",null,{"className":"text-xs text-zinc-500","children":[" · ","SynthScribe"]}]]}],["$","span",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2024"}]]}],["$","li","1",{"className":"avoid-break flex items-baseline justify-between gap-3 text-sm","children":[["$","span",null,{"className":"text-zinc-800","children":[["$","span",null,{"className":"font-medium","children":"Wolfond Scholarship"}],[["$","span",null,{"className":"text-zinc-500","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"University of Toronto"}]],["$","span",null,{"className":"text-xs text-zinc-500","children":[" · ","$$5,000 CAD"]}]]}],["$","span",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2023 to 2024"}]]}],["$","li","2",{"className":"avoid-break flex items-baseline justify-between gap-3 text-sm","children":[["$","span",null,{"className":"text-zinc-800","children":[["$","span",null,{"className":"font-medium","children":"DiDi Graduate Student Award in Computer Science"}],[["$","span",null,{"className":"text-zinc-500","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"University of Toronto"}]],["$","span",null,{"className":"text-xs text-zinc-500","children":[" · ","$$10,000 CAD, two years"]}]]}],["$","span",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2021 to 2023"}]]}],["$","li","3",{"className":"avoid-break flex items-baseline justify-between gap-3 text-sm","children":[["$","span",null,{"className":"text-zinc-800","children":[["$","span",null,{"className":"font-medium","children":"Apple AI/ML Fellowship Nominee"}],[["$","span",null,{"className":"text-zinc-500","children":" · "}],["$","span",null,{"className":"text-zinc-700","children":"University of Toronto"}]],["$","span",null,{"className":"text-xs text-zinc-500","children":[" · ","one of three at the school"]}]]}],["$","span",null,{"className":"text-xs text-zinc-500 tabular-nums whitespace-nowrap","children":"2022"}]]}]]}]]}],["$","section",null,{"className":"mb-6","children":[["$","h2",null,{"className":"font-serif text-sm tracking-[0.2em] uppercase text-zinc-800 border-b border-zinc-300 pb-1 mb-3","children":"Skills"}],["$","dl",null,{"className":"space-y-1.5","children":[["$","div","0",{"className":"grid grid-cols-[7rem_1fr] gap-3 text-sm avoid-break","children":[["$","dt",null,{"className":"text-xs font-semibold tracking-wider uppercase text-zinc-500 pt-0.5","children":"Research Areas"}],["$","dd",null,{"className":"text-zinc-800","children":"Human-AI Interaction · Interaction Design · Creativity Support Tools · Video & Audio Production · LLM Agents & UI Automation"}]]}],["$","div","1",{"className":"grid grid-cols-[7rem_1fr] gap-3 text-sm avoid-break","children":[["$","dt",null,{"className":"text-xs font-semibold tracking-wider uppercase text-zinc-500 pt-0.5","children":"Engineering"}],["$","dd",null,{"className":"text-zinc-800","children":"Python · TypeScript / React / Next.js / CSS · PyTorch · LLM tooling"}]]}]]}]]}]]}]]}],[["$","link","0",{"rel":"stylesheet","href":"/~bryanw/_next/static/css/57827a59dd5eb61a.css","precedence":"next","crossOrigin":"$undefined","nonce":"$undefined"}]],["$","$L5",null,{"children":"$L6"}]]}],{},null,false]},null,false]},null,false],["$","$1","h",{"children":[null,["$","$1","dli6iVHOys9Nix1vUOOHK",{"children":[["$","$L7",null,{"children":"$L8"}],["$","$L9",null,{"children":"$La"}],["$","meta",null,{"name":"next-size-adjust","content":""}]]}]]}],false]],"m":"$undefined","G":["$b","$undefined"],"s":false,"S":true}
a:[["$","meta","0",{"name":"viewport","content":"width=device-width, initial-scale=1"}]]
8:[["$","meta","0",{"charSet":"utf-8"}],["$","title","1",{"children":"Résumé · Bryan Wang, Ph.D."}],["$","meta","2",{"name":"description","content":"Résumé of Bryan Wang, Ph.D.."}],["$","link","3",{"rel":"icon","href":"/favicon.ico"}]]
6:null