Home>Scholars

  • Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
成果搜索
High Impact Results & Cited Count Trend for Year Keyword Cloud and Partner Relationship

Query:

学者姓名:兰诚栋

Refining:

Language

Submit

Clean All

Sort by:
Default
  • Default
  • Title
  • Year
  • WOS Cited Count
  • Impact factor
  • Ascending
  • Descending
< Page ,Total 5 >
GAN-based multi-view video coding with spatio-temporal EPI reconstruction SCIE
期刊论文 | 2025 , 132 | SIGNAL PROCESSING-IMAGE COMMUNICATION
Abstract&Keyword Cite Version(2)

Abstract :

The introduction of multiple viewpoints in video scenes inevitably increases the bitrates required for storage and transmission. To reduce bitrates, researchers have developed methods to skip intermediate viewpoints during compression and delivery, and ultimately reconstruct them using Side Information (SInfo). Typically, depth maps are used to construct SInfo. However, these methods suffer from reconstruction inaccuracies and inherently high bitrates. In this paper, we propose a novel multi-view video coding method that leverages the image generation capabilities of Generative Adversarial Network (GAN) to improve the reconstruction accuracy of SInfo. Additionally, we consider incorporating information from adjacent temporal and spatial viewpoints to further reduce SInfo redundancy. At the encoder, we construct a spatio-temporal Epipolar Plane Image (EPI) and further utilize a convolutional network to extract the latent code of a GAN as SInfo. At the decoder, we combine the SInfo and adjacent viewpoints to reconstruct intermediate views using the GAN generator. Specifically, we establish a joint encoder constraint for reconstruction cost and SInfo entropy to achieve an optimal trade-off between reconstruction quality and bitrate overhead. Experiments demonstrate the significant improvement in Rate-Distortion (RD) performance compared to state-of-the-art methods.

Keyword :

Epipolar plane image Epipolar plane image Generative adversarial network Generative adversarial network Latent code learning Latent code learning Multi-view video coding Multi-view video coding

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Lan, Chengdong , Yan, Hao , Luo, Cheng et al. GAN-based multi-view video coding with spatio-temporal EPI reconstruction [J]. | SIGNAL PROCESSING-IMAGE COMMUNICATION , 2025 , 132 .
MLA Lan, Chengdong et al. "GAN-based multi-view video coding with spatio-temporal EPI reconstruction" . | SIGNAL PROCESSING-IMAGE COMMUNICATION 132 (2025) .
APA Lan, Chengdong , Yan, Hao , Luo, Cheng , Zhao, Tiesong . GAN-based multi-view video coding with spatio-temporal EPI reconstruction . | SIGNAL PROCESSING-IMAGE COMMUNICATION , 2025 , 132 .
Export to NoteExpress RIS BibTex

Version :

GAN-based multi-view video coding with spatio-temporal EPI reconstruction EI
期刊论文 | 2025 , 132 | Signal Processing: Image Communication
GAN-based multi-view video coding with spatio-temporal EPI reconstruction Scopus
期刊论文 | 2025 , 132 | Signal Processing: Image Communication
Scale-Adaptive Asymmetric Sparse Variational AutoEncoder for Point Cloud Compression SCIE
期刊论文 | 2024 , 70 (3) , 884-894 | IEEE TRANSACTIONS ON BROADCASTING
Abstract&Keyword Cite Version(2)

Abstract :

Learning-based point cloud compression has achieved great success in Rate-Distortion (RD) efficiency. Existing methods usually utilize Variational AutoEncoder (VAE) network, which might lead to poor detail reconstruction and high computational complexity. To address these issues, we propose a Scale-adaptive Asymmetric Sparse Variational AutoEncoder (SAS-VAE) in this work. First, we develop an Asymmetric Multiscale Sparse Convolution (AMSC), which exploits multi-resolution branches to aggregate multiscale features at encoder, and excludes symmetric feature fusion branches to control the model complexity at decoder. Second, we design a Scale Adaptive Feature Refinement Structure (SAFRS) to adaptively adjust the number of Feature Refinement Modules (FRMs), thereby improving RD performance with an acceptable computational overhead. Third, we implement our framework with AMSC and SAFRS, and train it with an RD loss based on Fine-grained Weighted Binary Cross-Entropy (FWBCE) function. Experimental results on 8iVFB, Owlii, and MVUV datasets show that our method outperforms several popular methods, with a 90.0% time reduction and a 51.8% BD-BR saving compared with V-PCC. The code will be available soon at https://github.com/fancj2017/SAS-VAE.

Keyword :

asymmetric multiscale sparse convolution asymmetric multiscale sparse convolution Convolution Convolution Decoding Decoding Feature extraction Feature extraction Octrees Octrees Point cloud compression Point cloud compression Rate-distortion Rate-distortion scale adaptive feature refinement structure scale adaptive feature refinement structure Three-dimensional displays Three-dimensional displays variational autoencoder variational autoencoder

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Chen, Jian , Zhu, Yingtao , Huang, Wei et al. Scale-Adaptive Asymmetric Sparse Variational AutoEncoder for Point Cloud Compression [J]. | IEEE TRANSACTIONS ON BROADCASTING , 2024 , 70 (3) : 884-894 .
MLA Chen, Jian et al. "Scale-Adaptive Asymmetric Sparse Variational AutoEncoder for Point Cloud Compression" . | IEEE TRANSACTIONS ON BROADCASTING 70 . 3 (2024) : 884-894 .
APA Chen, Jian , Zhu, Yingtao , Huang, Wei , Lan, Chengdong , Zhao, Tiesong . Scale-Adaptive Asymmetric Sparse Variational AutoEncoder for Point Cloud Compression . | IEEE TRANSACTIONS ON BROADCASTING , 2024 , 70 (3) , 884-894 .
Export to NoteExpress RIS BibTex

Version :

Scale-Adaptive Asymmetric Sparse Variational AutoEncoder for Point Cloud Compression Scopus
期刊论文 | 2024 , 70 (3) , 884-894 | IEEE Transactions on Broadcasting
Scale-Adaptive Asymmetric Sparse Variational AutoEncoder for Point Cloud Compression EI
期刊论文 | 2024 , 70 (3) , 884-894 | IEEE Transactions on Broadcasting
DLKN: enhanced lightweight image super-resolution with dynamic large kernel network SCIE
期刊论文 | 2024 , 41 (5) , 3627-3644 | VISUAL COMPUTER
Abstract&Keyword Cite Version(4)

Abstract :

Convolutional neural networks are constrained in adaptively capturing information due to the use of fixed-size kernels. Although they provide a wide receptive field and achieve competitive performance with fewer parameters by using decomposed large kernels, they lack adaptability. Therefore, we propose the dynamic large kernel network (DLKN) for lightweight image super-resolution. Specifically, we design a basic convolutional block of feature aggregation groups, akin to the transformer architecture. It comprises a dynamic large kernel attention block and a local feature enhancement block that can adaptively utilize information. In our dynamic large kernel attention block, we decompose the large kernel convolution into kernels with different sizes and expansion rates. We then fuse their information for weight selection, dynamically adjusting the proportion of information from different receptive fields. The local feature enhancement block significantly improves local feature extraction with low parameter counts. It encourages interactions between local spatial features by decomposing the convolution into horizontally and vertically cascading kernels. Experimental results on benchmark datasets demonstrate that our proposed model achieves excellent performance in lightweight and performance-oriented super-resolution tasks. It successfully balances the relationship between performance and model complexity. The code is available at https://github.com/LyTinGiu/DLKN_SR.

Keyword :

CNN CNN Image processing Image processing Large kernel convolution Large kernel convolution Super-resolution Super-resolution

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Liu, YaTing , Lan, ChengDong , Feng, Wanjian . DLKN: enhanced lightweight image super-resolution with dynamic large kernel network [J]. | VISUAL COMPUTER , 2024 , 41 (5) : 3627-3644 .
MLA Liu, YaTing et al. "DLKN: enhanced lightweight image super-resolution with dynamic large kernel network" . | VISUAL COMPUTER 41 . 5 (2024) : 3627-3644 .
APA Liu, YaTing , Lan, ChengDong , Feng, Wanjian . DLKN: enhanced lightweight image super-resolution with dynamic large kernel network . | VISUAL COMPUTER , 2024 , 41 (5) , 3627-3644 .
Export to NoteExpress RIS BibTex

Version :

DLKN: enhanced lightweight image super-resolution with dynamic large kernel network Scopus
期刊论文 | 2025 , 41 (5) , 3627-3644 | Visual Computer
DLKN: enhanced lightweight image super-resolution with dynamic large kernel network SCIE
期刊论文 | 2025 , 41 (5) , 3627-3644 | VISUAL COMPUTER
DLKN: enhanced lightweight image super-resolution with dynamic large kernel network EI
期刊论文 | 2025 , 41 (5) , 3627-3644 | Visual Computer
DLKN: enhanced lightweight image super-resolution with dynamic large kernel network Scopus
期刊论文 | 2024 | Visual Computer
Adaptive Streaming of Stereoscopic Panoramic Video Based on Reinforcement Learning EI CSCD PKU
期刊论文 | 2022 , 44 (4) , 1461-1468 | Journal of Electronics and Information Technology
Abstract&Keyword Cite

Abstract :

Currently, an effective stream adaptation method for stereo panoramic video transmission is missing. However, the traditional panoramic video adaptive streaming strategy for transmitting binocular stereo panoramic video suffers from the problem of doubling the transmission data and requiring huge bandwidth. A multi-agent reinforcement learning based stereo panoramic video asymmetric transmission adaptive streaming method is proposed in this paper to cope with the limited bandwidth and fluctuation of network bandwidth in real time. First, due to the human eye's preference for the saliency regions of video, each tile in the left and right viewpoints of stereoscopic video contributes differently to the perceptual quality, and a tiles-based method for predicting the watching probability of left and right viewpoint is proposed. Second, a multi-agent reinforcement learning framework based on policy-value (Actor-Critic) is designed for joint rate control of left and right viewpoints. Finally, a reasonable reward function is designed based on the model structure and the principle of binocular suppression. The experimental results show that the proposed method is more suitable for tiles-based stereo panoramic video transmission than the traditional self-adaptive stream transmission strategy. A novel approach is proposed for stereo panoramic video joint rate control and user Quality of Experience (QoE) improvement under limited bandwidth. © 2022, Science Press. All right reserved.

Keyword :

Bandwidth Bandwidth Image communication systems Image communication systems Multi agent systems Multi agent systems Quality control Quality control Quality of service Quality of service Reinforcement learning Reinforcement learning Stereo image processing Stereo image processing Video streaming Video streaming

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Lan, Chengdong , Rao, Yingjie , Song, Caixia et al. Adaptive Streaming of Stereoscopic Panoramic Video Based on Reinforcement Learning [J]. | Journal of Electronics and Information Technology , 2022 , 44 (4) : 1461-1468 .
MLA Lan, Chengdong et al. "Adaptive Streaming of Stereoscopic Panoramic Video Based on Reinforcement Learning" . | Journal of Electronics and Information Technology 44 . 4 (2022) : 1461-1468 .
APA Lan, Chengdong , Rao, Yingjie , Song, Caixia , Chen, Jian . Adaptive Streaming of Stereoscopic Panoramic Video Based on Reinforcement Learning . | Journal of Electronics and Information Technology , 2022 , 44 (4) , 1461-1468 .
Export to NoteExpress RIS BibTex

Version :

多视点稀疏测量的图像绘制方法 CSCD PKU
期刊论文 | 2021 , 47 (4) , 882-890 | 自动化学报
Abstract&Keyword Cite Version(1)

Abstract :

为了减少所需采集的视频数据量,基于图像绘制(Image-based rendering,IBR)的前沿方法将稠密视点信息映射成压缩感知框架中的原始信号,并将稀疏视点图像作为随机测量值,但低维测量信号由所有稠密视点信息线性组合而成,而稀疏视点图像仅仅来源于部分视点信息,导致稀疏视点采集的图像与低维测量信号不一致.本文提出利用间隔采样矩阵消除测量信号与稀疏视点图像位置之间的差异,进而通过约束由测量矩阵和基函数构成的传感矩阵尽量满足有限等距性,使得能够获得原始信号的唯一精确解.仿真实验结果表明,相比于前沿方法,本文提出的方法对于不同复杂程度的场景重建都提高了主客观质量.

Keyword :

压缩感知 压缩感知 基于图像的绘制 基于图像的绘制 多视点图像重建 多视点图像重建 极平面图像 极平面图像

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 兰诚栋 , 林宇鹏 , 方大锐 et al. 多视点稀疏测量的图像绘制方法 [J]. | 自动化学报 , 2021 , 47 (4) : 882-890 .
MLA 兰诚栋 et al. "多视点稀疏测量的图像绘制方法" . | 自动化学报 47 . 4 (2021) : 882-890 .
APA 兰诚栋 , 林宇鹏 , 方大锐 , 陈建 . 多视点稀疏测量的图像绘制方法 . | 自动化学报 , 2021 , 47 (4) , 882-890 .
Export to NoteExpress RIS BibTex

Version :

Multi-view Sparse Measurement for Image-based Rendering Method EI PKU
期刊论文 | 2021 , 47 (4) , 882-890 | Acta Automatica Sinica
多视点稀疏测量的图像绘制方法 CSCD PKU
期刊论文 | 2021 , 47 (04) , 882-890 | 自动化学报
Abstract&Keyword Cite

Abstract :

为了减少所需采集的视频数据量,基于图像绘制(Image-based rendering, IBR)的前沿方法将稠密视点信息映射成压缩感知框架中的原始信号,并将稀疏视点图像作为随机测量值,但低维测量信号由所有稠密视点信息线性组合而成,而稀疏视点图像仅仅来源于部分视点信息,导致稀疏视点采集的图像与低维测量信号不一致.本文提出利用间隔采样矩阵消除测量信号与稀疏视点图像位置之间的差异,进而通过约束由测量矩阵和基函数构成的传感矩阵尽量满足有限等距性,使得能够获得原始信号的唯一精确解.仿真实验结果表明,相比于前沿方法,本文提出的方法对于不同复杂程度的场景重建都提高了主客观质量.

Keyword :

压缩感知 压缩感知 基于图像的绘制 基于图像的绘制 多视点图像重建 多视点图像重建 极平面图像 极平面图像

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 兰诚栋 , 林宇鹏 , 方大锐 et al. 多视点稀疏测量的图像绘制方法 [J]. | 自动化学报 , 2021 , 47 (04) : 882-890 .
MLA 兰诚栋 et al. "多视点稀疏测量的图像绘制方法" . | 自动化学报 47 . 04 (2021) : 882-890 .
APA 兰诚栋 , 林宇鹏 , 方大锐 , 陈建 . 多视点稀疏测量的图像绘制方法 . | 自动化学报 , 2021 , 47 (04) , 882-890 .
Export to NoteExpress RIS BibTex

Version :

Reconstruction of Multi-view Video Based on GAN CPCI-S
会议论文 | 2018 , 11165 , 618-629 | 19th Pacific-Rim Conference on Multimedia (PCM)
Abstract&Keyword Cite Version(2)

Abstract :

There is a huge amount of data in multi-view video which brings enormous challenges to the compression, storage, and transmission of video data. Transmitting part of the viewpoint information is a prior solution to reconstruct the original multi-viewpoint information. They are all based on pixel matching to obtain the correlation between adjacent viewpoint images. However, pixels cannot express the invariability of image features and are susceptible to noise. Therefore, in order to overcome the above problems, the VGG network is used to extract the high-dimensional features between the images, indicating the relevance of the adjacent images. The GAN is further used to more accurately generate virtual viewpoint images. We extract the lines at the same positions of the viewpoints as local areas for image merging and input the local images into the network. In the reconstruction viewpoint, we generate a local image of a dense viewpoint through the GAN network. Experiments on multiple test sequences show that the proposed method has a 0.2-0.8-dB PSNR and 0.15-0.61 MOS improvement over the traditional method.

Keyword :

EPI EPI Hybrid resolution Hybrid resolution Multi-view video Multi-view video SRGAN SRGAN Virtual view reconstruction Virtual view reconstruction

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Li, Song , Lan, Chengdong , Zhao, Tiesong . Reconstruction of Multi-view Video Based on GAN [C] . 2018 : 618-629 .
MLA Li, Song et al. "Reconstruction of Multi-view Video Based on GAN" . (2018) : 618-629 .
APA Li, Song , Lan, Chengdong , Zhao, Tiesong . Reconstruction of Multi-view Video Based on GAN . (2018) : 618-629 .
Export to NoteExpress RIS BibTex

Version :

Reconstruction of multi-view video based on gan Scopus
会议论文 | 2018 , 11165 LNCS , 618-629 | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Reconstruction of multi-view video based on gan EI
会议论文 | 2018 , 11165 LNCS , 618-629
全景纵向漫游中极线匹配的置信传播算法 CSCD PKU
期刊论文 | 2018 , 30 (3) , 400-407 | 计算机辅助设计与图形学学报
Abstract&Keyword Cite Version(2)

Abstract :

深度信息的获取是全景纵向漫游的关键基础,为了提高前后场景图像匹配的精确度,提出一种极线匹配的置信传播算法.首先根据对极几何原理,以图像中心点为基准全方向发散构建出前后场景图像的对极线;其次利用对极线路径信息在匹配代价函数中增加垂直方向的匹配代价分量,并用置信传播算法生成视差图;最后通过前后场景图像的几何关系构建视差与深度的计算模型,从而获取深度信息.实验结果表明,与局部优化算法相比,该算法深度图的结构相似性平均提高了0.22,峰值信噪比平均提高了24%,对于前后场景图像深度信息的获取具有更好的效果.

Keyword :

代价函数 代价函数 全景纵向漫游 全景纵向漫游 极线匹配 极线匹配 深度图 深度图 置信传播 置信传播

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 孙强强 , 兰诚栋 , 陈康杰 et al. 全景纵向漫游中极线匹配的置信传播算法 [J]. | 计算机辅助设计与图形学学报 , 2018 , 30 (3) : 400-407 .
MLA 孙强强 et al. "全景纵向漫游中极线匹配的置信传播算法" . | 计算机辅助设计与图形学学报 30 . 3 (2018) : 400-407 .
APA 孙强强 , 兰诚栋 , 陈康杰 , 方大锐 . 全景纵向漫游中极线匹配的置信传播算法 . | 计算机辅助设计与图形学学报 , 2018 , 30 (3) , 400-407 .
Export to NoteExpress RIS BibTex

Version :

Belief Propagation Algorithm for Epipolar Matching in Panoramic Longitudinal Roamin EI CSCD PKU
期刊论文 | 2018 , 30 (3) , 400-407 | Journal of Computer-Aided Design and Computer Graphics
Belief Propagation Algorithm for Epipolar Matching in Panoramic Longitudinal Roamin [全景纵向漫游中极线匹配的置信传播算法] Scopus CSCD PKU
期刊论文 | 2018 , 30 (3) , 400-407 | Journal of Computer-Aided Design and Computer Graphics
基于多相机系统标定的全景拼接算法研究
期刊论文 | 2018 , (3) , 37-40 | 信息通信
Abstract&Keyword Cite Version(2)

Abstract :

全景拼接技术是当前全景技术的重点,而多相机系统被越来越多的应用于全景技术中.文章提出了一种基于多相机系统标定的全景拼接方法,利用通过标定得到的相机系统相关参数,将多相机系统拍摄得到的图像进行畸变矫正与旋转变换处理,以提高全景拼接的效果.使用基于特征描述符的标定板对多相机系统进行标定,获得相机的内部参数和外部参数;利用Mercator投影,结合标定得到的内部参数对鱼眼图像进行畸变矫正,外部参数则用于对矫正后的图像进行旋转变换处理,使它们在同一世界坐标系下;用Harris算法对相邻待拼接图像的特征点进行检测匹配,进而拼接成全景图.通过实验验证,该全景拼接算法具有不错的拼接精度,没有出现误匹配的情况.

Keyword :

全景拼接 全景拼接 多相机系统标定 多相机系统标定 旋转变换 旋转变换 畸变矫正 畸变矫正

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 林华琦 , 兰诚栋 . 基于多相机系统标定的全景拼接算法研究 [J]. | 信息通信 , 2018 , (3) : 37-40 .
MLA 林华琦 et al. "基于多相机系统标定的全景拼接算法研究" . | 信息通信 3 (2018) : 37-40 .
APA 林华琦 , 兰诚栋 . 基于多相机系统标定的全景拼接算法研究 . | 信息通信 , 2018 , (3) , 37-40 .
Export to NoteExpress RIS BibTex

Version :

基于多相机系统标定的全景拼接算法研究 CQVIP
期刊论文 | 2018 , 0 (3) , 37-40 | 信息通信
基于多相机系统标定的全景拼接算法研究
期刊论文 | 2018 , (03) , 37-40 | 信息通信
全景纵向漫游中极线匹配的置信传播算法 CSCD PKU
期刊论文 | 2018 , 30 (03) , 400-407 | 计算机辅助设计与图形学学报
Abstract&Keyword Cite Version(1)

Abstract :

深度信息的获取是全景纵向漫游的关键基础,为了提高前后场景图像匹配的精确度,提出一种极线匹配的置信传播算法.首先根据对极几何原理,以图像中心点为基准全方向发散构建出前后场景图像的对极线;其次利用对极线路径信息在匹配代价函数中增加垂直方向的匹配代价分量,并用置信传播算法生成视差图;最后通过前后场景图像的几何关系构建视差与深度的计算模型,从而获取深度信息.实验结果表明,与局部优化算法相比,该算法深度图的结构相似性平均提高了0.22,峰值信噪比平均提高了24%,对于前后场景图像深度信息的获取具有更好的效果.

Keyword :

代价函数 代价函数 全景纵向漫游 全景纵向漫游 极线匹配 极线匹配 深度图 深度图 置信传播 置信传播

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 孙强强 , 兰诚栋 , 陈康杰 et al. 全景纵向漫游中极线匹配的置信传播算法 [J]. | 计算机辅助设计与图形学学报 , 2018 , 30 (03) : 400-407 .
MLA 孙强强 et al. "全景纵向漫游中极线匹配的置信传播算法" . | 计算机辅助设计与图形学学报 30 . 03 (2018) : 400-407 .
APA 孙强强 , 兰诚栋 , 陈康杰 , 方大锐 . 全景纵向漫游中极线匹配的置信传播算法 . | 计算机辅助设计与图形学学报 , 2018 , 30 (03) , 400-407 .
Export to NoteExpress RIS BibTex

Version :

全景纵向漫游中极线匹配的置信传播算法 CQVIP CSCD PKU
期刊论文 | 2018 , 30 (3) , 400-407 | 计算机辅助设计与图形学学报
10| 20| 50 per page
< Page ,Total 5 >

Export

Results:

Selected

to

Format:
Online/Total:62/10077771
Address:FZU Library(No.2 Xuyuan Road, Fuzhou, Fujian, PRC Post Code:350116) Contact Us:0591-22865326
Copyright:FZU Library Technical Support:Beijing Aegean Software Co., Ltd. 闽ICP备05005463号-1