Development of model based on clock gene expression of human hair follicle cells to estimate circadian time
- Authors
- Lee, Taek; Cho, Chul-Hyun; Kim, Woon Ryoung; Moon, Joung Ho; Kim, Soojin; Geum, Dongho; In, Hoh Peter; Lee, Heon-Jeong
- Issue Date
- 2-7월-2020
- Publisher
- TAYLOR & FRANCIS INC
- Keywords
- Circadian clock; circadian genes; hair follicle; machine learning; circadian time estimation
- Citation
- CHRONOBIOLOGY INTERNATIONAL, v.37, no.7, pp.993 - 1001
- Indexed
- SCIE
SCOPUS
- Journal Title
- CHRONOBIOLOGY INTERNATIONAL
- Volume
- 37
- Number
- 7
- Start Page
- 993
- End Page
- 1001
- URI
- https://scholar.korea.ac.kr/handle/2021.sw.korea/54443
- DOI
- 10.1080/07420528.2020.1777150
- ISSN
- 0742-0528
- Abstract
- Considering the effects of circadian misalignment on human pathophysiology and behavior, it is important to be able to detect an individual's endogenous circadian time. We developed an endogenous Clock Estimation Model (eCEM) based on a machine learning process using the expression of 10 circadian genes. Hair follicle cells were collected from 18 healthy subjects at 08:00, 11:00, 15:00, 19:00, and 23:00 h for two consecutive days, and the expression patterns of 10 circadian genes were obtained. The eCEM was designed using the inverse form of the circadian gene rhythm function (i.e., Circadian Time = F(gene)), and the accuracy of eCEM was evaluated by leave-one-out cross-validation (LOOCV). As a result, six genes (PER1, PER3, CLOCK, CRY2, NPAS2, andNR1D2)were selected as the best model, and the error range between actual and predicted time was 3.24 h. The eCEM is simple and applicable in that a single time-point sampling of hair follicle cells at any time of the day is sufficient to estimate the endogenous circadian time.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Graduate School > Department of Biomedical Sciences > 1. Journal Articles
- Graduate School > Department of Computer Science and Engineering > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.