Please use this identifier to cite or link to this item:
https://elib.bsu.by/handle/123456789/10020Full metadata record
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Lisitsa, Y. | - |
| dc.contributor.author | Yatskou, M. | - |
| dc.contributor.author | Apanasovich, N. | - |
| dc.contributor.author | Apanasovich, V. | - |
| dc.date.accessioned | 2012-05-25T08:11:16Z | - |
| dc.date.available | 2012-05-25T08:11:16Z | - |
| dc.date.issued | 2011 | - |
| dc.identifier.citation | Международный конгресс по информатике: информационные системы и технологии: материалы международного научного конгресса 31 окт. – 3 нояб. 2011 г. : в 2 ч. Ч. 2. – Минск: БГУ, 2011. – C. 242-247. | ru |
| dc.identifier.isbn | 978-985-518-564-3 | - |
| dc.identifier.uri | http://elib.bsu.by/handle/123456789/10020 | - |
| dc.description | Секция 9. Распознавание образов, информационные системы управления | ru |
| dc.description.abstract | Fluorescent microscopy imaging is an efficient technique allowing estimating about protein expression in cells. A critical issue is image analysis. The complexity of investigated cell systems hampers application of standard image processing methods. We have developed an image analysis approach for the nuclear segmentation, that is (i) fully-automated, (ii) high-throughput, and (iii) has the improved accuracy of segmentation. The developed approach exploits the morphological information about nuclei in the image and pixel intensity values. | ru |
| dc.language.iso | ru | ru |
| dc.publisher | БГУ | ru |
| dc.subject | ЭБ БГУ::ОБЩЕСТВЕННЫЕ НАУКИ::Информатика | ru |
| dc.title | A fully-automated nuclear segmentation approach for analysis of cell fluorescent images | ru |
| dc.type | Article | ru |
| Appears in Collections: | 2011. Международный конгресс по информатике : информационные системы и технологии. Часть 2. | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| 50 Lisitsa.pdf | 549,63 kB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

