-
Notifications
You must be signed in to change notification settings - Fork 1
Expand file tree
/
Copy pathindex.html
More file actions
107 lines (96 loc) · 5.2 KB
/
index.html
File metadata and controls
107 lines (96 loc) · 5.2 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
<!DYPE HTML>
<html lang="en"><head><meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
<title>Ziyi Li</title>
<meta name="author" content="Ziyi Li">
<meta name="viewport" content="width=device-width, initial-scale=1">
<link rel="stylesheet" type="text/css" href="stylesheet.css">
</head>
<body>
<table style="width:100%;max-width:800px;border:0px;border-spacing:0px;border-collapse:separate;margin-right:auto;margin-left:auto;"><tbody>
<tr style="padding:0px">
<td style="padding:0px">
<table style="width:100%;border:0px;border-spacing:0px;border-collapse:separate;margin-right:auto;margin-left:auto;"><tbody>
<tr style="padding:0px">
<td style="padding:2.5%;width:63%;vertical-align:middle">
<p style="text-align:center">
<name> Ziyi Li </name>
</p>
<p>I am a master student working on open-world representation learning and image super-resolution at <a href="https://mediabrain.sjtu.edu.cn/">Shanghai Jiao Tong University</a>.
I received my bachelor degree from <a href='https://www.sjtu.edu.cn/'> Shanghai Jiao Tong University (SJTU)</a> in June 2021.
</p>
<p>
My current research interest is in open-world representation learning.
</p>
<p style="text-align:center">
<a href="599lzy@sjtu.edu.cn">Email</a>  / 
<a href="https://github.com/Lipurple">Github</a>
</p>
</td>
<td style="padding:2.5%;width:40%;max-width:40%">
<a href="image/lzy.jpg"><img style="width:100%;max-width:100%" alt="profile photo" src="image/liziyi.jpg" class="hoverZoomLink"></a>
</td>
</tr>
</tbody></table>
<br />
<table style="width:100%;border:0px;border-spacing:0px;border-collapse:separate;margin-right:auto;margin-left:auto;"><tbody>
<heading><b></bold>Research</b></heading>
<tr onmouseout="ld_stop()" onmouseover="ld_start()">
<td style="padding:0px;width:25%;vertical-align:middle">
<img src="image/grounded.png" width="320" height="100" style="border-style: none">
</td>
<td style="padding:10px;width:75%;vertical-align:middle">
<a href="https://lipurple.github.io/Grounded_Diffusion/">
<papertitle>Guiding Text-to-Image Diffusion Model Towards Grounded Generation</papertitle>
</a>
<br>
<strong>Ziyi Li* </strong>,
<a>Qinye Zhou* </a>,
<a href="https://cmic.sjtu.edu.cn/CN/show.aspx?info_lb=35&info_id=1341&flag=35">Xiaoyun Zhang </a>,
<a href="https://mediabrain.sjtu.edu.cn/yazhang/">Ya Zhang </a>,
<a href="https://mediabrain.sjtu.edu.cn/members/">Yanfeng Wang </a>,
<a href="https://weidixie.github.io/">Weidi Xie </a>
<br>
<em> International Conference on Computer Vision (ICCV)</em>, 2023
<!-- <em>arXiv</em>, 2023 (under review) -->
<p>we propose to augment a pre-trained text-to-image diffusion model with the ability of open-vocabulary objects grounding, i.e., simultaneously generating images and segmentation masks for the corresponding visual entities described in the text prompt.
</p>
</td>
</tr>
<tr onmouseout="ld_stop()" onmouseover="ld_start()">
<td style="padding:0px;width:40%;vertical-align:middle">
<img src="image/ARIS.png" width="320" height="160" style="border-style: none">
</td>
<td style="padding:10px;width:60%;vertical-align:middle">
<a href="https://lipurple.github.io/ARIS_Webpage//">
<papertitle>A Simple Plugin for Transforming Images to Arbitrary Scales</papertitle>
</a>
<br>
<a>Qinye Zhou* </a>,
<strong>Ziyi Li* </strong>,
<a href="https://weidixie.github.io/">Weidi Xie </a>,
<a href="https://cmic.sjtu.edu.cn/CN/show.aspx?info_lb=35&info_id=1341&flag=35">Xiaoyun Zhang </a>,
<a href="https://mediabrain.sjtu.edu.cn/members/">Yanfeng Wang </a>,
<a href="https://mediabrain.sjtu.edu.cn/yazhang/">Ya Zhang</a>
<br>
<em> British Machine Vision Conference (BMVC)</em>, 2022
<p>we propose to develop a general <strong>plugin</strong> that can be inserted into existing super-resolution models, conveniently augmenting their ability towards <strong>A</strong>rbitrary <strong>R</strong>esolution <strong>I</strong>mage <strong>S</strong>caling, thus termed <strong>ARIS</strong>.</p>
</td>
</tr>
</tbody></table>
<table style="width:100%;border:0px;border-spacing:0px;border-collapse:separate;margin-right:auto;margin-left:auto;"><tbody>
<tr>
<td style="padding:0px">
<br>
<p style="text-align:right;font-size:small;">
Based on a template by <a href="https://github.com/jonbarron/jonbarron_website">Jon Barron</a>.
<br>
</p>
</td>
</tr>
</tbody></table>
</td>
</tr>
</table>
</body>
</html>
<!--for page update-->