| dc.citation.conferencePlace |
KO |
- |
| dc.citation.endPage |
810 |
- |
| dc.citation.startPage |
807 |
- |
| dc.citation.title |
2022년 대한전자공학회 하계종합학술대회 |
- |
| dc.contributor.author |
정주은 |
- |
| dc.contributor.author |
이규호 |
- |
| dc.date.accessioned |
2024-01-31T20:09:29Z |
- |
| dc.date.available |
2024-01-31T20:09:29Z |
- |
| dc.date.created |
2022-08-04 |
- |
| dc.date.issued |
2022-06-30 |
- |
| dc.description.abstract |
Semantic segmentation is one of the most fundamental perception tasks for Autonomous Electric vehicle (AEV). It provides an overall understanding of the driving environment, including road and pedestrians. Its high computational cost with high-resolution images makes real-time implementation difficult in time-critical and resource-constrained AEV. To resolve this issue, this paper proposes a Depth-fused Trilateral Network (DTN) with dilated convolution and depthwise separable convolution that reduces 90% of the overall computation of baseline network[1] and achieves 94.73% MaxF on KITTI Road dataset and 58.67% mIOU on Cityscape 7 dataset. |
- |
| dc.identifier.bibliographicCitation |
2022년 대한전자공학회 하계종합학술대회, pp.807 - 810 |
- |
| dc.identifier.uri |
https://scholarworks.unist.ac.kr/handle/201301/75755 |
- |
| dc.language |
한국어 |
- |
| dc.publisher |
대한전자공학회 |
- |
| dc.title |
깊이 정보를 활용한 자율주행을 위한 실시간 의미론적 영상 분할 네트워크 |
- |
| dc.type |
Conference Paper |
- |
| dc.date.conferenceDate |
2022-06-29 |
- |