共查询到20条相似文献,搜索用时 31 毫秒
1.
Chenling Xiong Katherina C. Chua Tore B. Stage Josefina Priotti Jeffrey Kim Anne AltmanMerino Daniel Chan Krishna Saraf Amanda Canato Ferracini Faranak Fattahi Deanna L. Kroetz 《CTS Clinical and Translational Science》2021,14(2):568
Chemotherapy‐induced peripheral neuropathy (CIPN) is a dose‐limiting adverse event associated with treatment with paclitaxel and other chemotherapeutic agents. The prevention and treatment of CIPN are limited by a lack of understanding of the molecular mechanisms underlying this toxicity. In the current study, a human induced pluripotent stem cell–derived sensory neuron (iPSC‐SN) model was developed for the study of chemotherapy‐induced neurotoxicity. The iPSC‐SNs express proteins characteristic of nociceptor, mechanoreceptor, and proprioceptor sensory neurons and show Ca2+ influx in response to capsaicin, α,β‐meATP, and glutamate. The iPSC‐SNs are relatively resistant to the cytotoxic effects of paclitaxel, with half‐maximal inhibitory concentration (IC50) values of 38.1 µM (95% confidence interval (CI) 22.9–70.9 µM) for 48‐hour exposure and 9.3 µM (95% CI 5.7–16.5 µM) for 72‐hour treatment. Paclitaxel causes dose‐dependent and time‐dependent changes in neurite network complexity detected by βIII‐tubulin staining and high content imaging. The IC50 for paclitaxel reduction of neurite area was 1.4 µM (95% CI 0.3–16.9 µM) for 48‐hour exposure and 0.6 µM (95% CI 0.09–9.9 µM) for 72‐hour exposure. Decreased mitochondrial membrane potential, slower movement of mitochondria down the neurites, and changes in glutamate‐induced neuronal excitability were also observed with paclitaxel exposure. The iPSC‐SNs were also sensitive to docetaxel, vincristine, and bortezomib. Collectively, these data support the use of iPSC‐SNs for detailed mechanistic investigations of genes and pathways implicated in chemotherapy‐induced neurotoxicity and the identification of novel therapeutic approaches for its prevention and treatment. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
2.
Wendy D. Woodley Wen Yue Didier R. Morel Audrey Lainesse Ronald J. Pettis Natasha G. Bolick 《CTS Clinical and Translational Science》2021,14(3):859
An investigational wearable injector (WI), the BD Libertas Wearable Injector (BD Libertas is a trademark of Becton, Dickinson and Company), was evaluated in an early feasibility clinical study for functional performance, tissue effects, subject tolerability, and acceptability of 5 mL, non‐Newtonian ~ 8 cP subcutaneous placebo injections in 52 healthy adult subjects of 2 age groups (18–64 years and ≥ 65 years). Randomized WI subcutaneous injections (n = 208, 4/subject) were delivered to the right and left abdomen and thigh of each subject, 50% (1 thigh and 1 abdomen) with a defined movement sequence during injection. Injector functional performance was documented. Deposition was qualified and quantified with ultrasound. Tissue effects and tolerability (pain) were monitored through 24 hours with corresponding acceptability questionnaires administered through 72 hours. WI (n = 205) automatically inserted the needle, delivered 5 mL ± 5% in 5.42 minutes (SD 0.74) and retracted. Depots were entirely (93.2%) or predominantly (5.4%) localized within the target subcutaneous tissue. Slight to moderate wheals (63.9%) and erythema (75.1%) were observed with ≥ 50% resolution within 30–60 minutes. Subject pain (100 mm Visual Analog Scale) peaked mid‐injection (mean 9.1 mm, SD 13.4) and rapidly resolved within 30 minutes (mean 0.4 mm, SD 2.6). Subjects’ peak pain (≥ 90.2%), injection site appearance (≥ 92.2%) and injector wear, size, and removal (≥ 92.1%) were acceptable (Likert responses) with 100% likely to use the injector if prescribed. Injection site preference was divided between none (46%), abdomen (25%), or thigh (26.9%). The investigational WI successfully delivered 5 mL viscous subcutaneous injections. Tissue effects and pain were transient, well‐tolerated and acceptable. Neither injection site, movement or subject age affected injector functional performance or subject pain and acceptability. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
3.
Tina Hrbelt Anna Lena Kahl Frederike Kolbe Susann Hetze Benjamin Wilde Oliver Witzke Manfred Schedlowski 《CTS Clinical and Translational Science》2020,13(6):1251
The rapamycin analogue everolimus (EVR) is a potent inhibitor of the mammalian target of rapamycin (mTOR) and clinically used to prevent allograft rejections as well as tumor growth. The pharmacokinetic and immunosuppressive efficacy of EVR have been extensively reported in patient populations and in vitro studies. However, dose‐dependent ex vivo effects upon acute EVR administration in healthy volunteers are rare. Moreover, immunosuppressive drugs are associated with neuroendocrine changes and psychological disturbances. It is largely unknown so far whether and to what extend EVR affects neuroendocrine functions, mood, and anxiety in healthy individuals. Thus, in the present study, we analyzed the effects of three different clinically applied EVR doses (1.5, 2.25, and 3 mg) orally administered 4 times in a 12‐hour cycle to healthy male volunteers on immunological, neuroendocrine, and psychological parameters. We observed that oral intake of medium (2.25 mg) and high doses (3 mg) of EVR efficiently suppressed T cell proliferation as well as IL‐10 cytokine production in ex vivo mitogen‐stimulated peripheral blood mononuclear cell. Further, acute low (1.5 mg) and medium (2.25 mg) EVR administration increased state anxiety levels accompanied by significantly elevated noradrenaline (NA) concentrations. In contrast, high‐dose EVR significantly reduced plasma and saliva cortisol as well as NA levels and perceived state anxiety. Hence, these data confirm the acute immunosuppressive effects of the mTOR inhibitor EVR and provide evidence for EVR‐induced alterations in neuroendocrine parameters and behavior under physiological conditions in healthy volunteers. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
4.
Ravi Shankar P. Singh Vivek Pradhan Erika S. Roberts Matthew Scaramozza Elizabeth Kieras Jeremy D. Gale Elena Peeva Michael S. Vincent Anindita Banerjee Andrew Fensome Martin E. Dowty Peter Winkle Christopher Tehlirian 《CTS Clinical and Translational Science》2021,14(2):671
Selective inhibition of tyrosine kinase 2 (TYK2) may offer therapeutic promise in inflammatory conditions, with its role in downstream pro‐inflammatory cytokine signaling. In this first‐in‐human study, we evaluated the safety, tolerability, and pharmacokinetics (PK) of a novel TYK2 inhibitor, PF‐06826647, in healthy participants. This phase I, randomized, double‐blind, placebo‐controlled, parallel‐group study included two treatment periods (single ascending dose (SAD) and multiple ascending dose (MAD)) in healthy participants and a cohort of healthy Japanese participants receiving 400 mg q.d. or placebo in the MAD period (). Participants were randomly assigned to PF‐06826647 or placebo (3:1). Participants received a single oral study drug dose of 3, 10, 30, 100, 200, 400, or 1,600 mg (SAD period), then 30, 100, 400, or 1,200 mg q.d. or 200 mg b.i.d. for 10 days (MAD period). Safety (adverse events (AEs), vital signs, and clinical laboratory parameters), tolerability, and PK were assessed. Overall, 69 participants were randomized to treatment, including six Japanese participants. No deaths, serious AEs, severe AEs, or AEs leading to dose reduction or temporary/permanent discontinuation were observed. All AEs were mild in severity. No clinically relevant laboratory abnormalities or changes in vital signs were detected. PF‐06826647 was rapidly absorbed with a median time to maximum plasma concentration of 2 hours in a fasted state, with modest accumulation (< 1.5‐fold) after multiple dosing and low urinary recovery. PF‐06826647 was well‐tolerated, with an acceptable safety profile for doses up to 1,200 mg q.d. for 10 days, supporting further testing in patients. Study Highlights NCT03210961
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
5.
6.
Nicolas Hohmann Farastuk Bozorgmehr Petros Christopoulos Gerd Mikus Antje Blank Jürgen Burhenne Michael Thomas Walter E. Haefeli 《CTS Clinical and Translational Science》2021,14(2):487
The inhibitor of anaplastic lymphoma kinase (ALK) crizotinib significantly increases survival in patients with ALK‐positive non‐small cell lung cancer (NSCLC). When evaluating crizotinib pharmacokinetics (PKs) in patients taking the standard flat oral dose of 250 mg b.i.d., interindividual PK variability is substantial and patient survival is lower in the quartile with the lowest steady‐state trough plasma concentrations (Cmin,ss), suggesting that concentrations should be monitored and doses individualized. We investigated whether the CYP3A inhibitor cobicistat increases Cmin,ss of the CYP3A substrate crizotinib in patients with low exposure. Patients with ALK‐positive NSCLC of our outpatient clinic treated with crizotinib were enrolled in a phase I trial (EudraCT 2016‐002187‐14, DRKS00012360) if crizotinib Cmin,ss was below 310 ng/mL and treated with cobicistat for 14 days. Crizotinib plasma concentration profiles were established before and after a 14‐day co‐administration of cobicistat to construct the area under the plasma concentration‐time curve in the dosing interval from zero to 12 hours (AUC0–12). Patients were also monitored for adverse events by physical examination, laboratory tests, and 12‐lead echocardiogram. Enrolment was prematurely stopped because of the approval of alectinib, a next‐generation ALK‐inhibitor with superior efficacy. In the only patient enrolled, cobicistat increased Cmin,ss from 158 ng/mL (before cobicistat) to 308 ng/mL (day 8) and 417 ng/mL (day 14 on cobicistat), concurrently the AUC0–12 increased by 78% from 2,210 ng/mL*h to 3,925 ng/mL*h. Neither safety signals nor serious adverse events occurred. Pharmacoenhancement with cobicistat as an alternative for dose individualisation for patients with NSCLC with low crizotinib exposure appears to be safe and is cost‐effective and feasible. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
7.
Xiaoyu Yan Xu Steven Xu Katja C. Weisel MariaVictoria Mateos Pieter Sonneveld Meletios A. Dimopoulos Saad Zafar Usmani Nizar J. Bahlis Thomas Puchalski Jon Ukropec Kevin Bellew Qi Ming Steven Sun Honghui Zhou 《CTS Clinical and Translational Science》2020,13(6):1345
This study aimed to predict long‐term progression‐free survival (PFS) using early M‐protein dynamic measurements in patients with relapsed/refractory multiple myeloma (MM). The PFS was modeled based on dynamic M‐protein data from two phase III studies, POLLUX and CASTOR, which included 569 and 498 patients with relapsed/refractory MM, respectively. Both studies compared active controls (lenalidomide and dexamethasone, and bortezomib and dexamethasone, respectively) alone vs. in combination with daratumumab. Three M‐protein dynamic features from the longitudinal M‐protein data were evaluated up to different time cutoffs (1, 2, 3, and 6 months). The abilities of early M‐protein dynamic measurements to predict the PFS were evaluated using Cox proportional hazards survival models. Both univariate and multivariable analyses suggest that maximum reduction of M‐protein (i.e., depth of response) was the most predictive of PFS. Despite the statistical significance, the baseline covariates provided very limited predictive value regarding the treatment effect of daratumumab. However, M‐protein dynamic features obtained within the first 2 months reasonably predicted PFS and the associated treatment effect of daratumumab. Specifically, the areas under the time‐varying receiver operating characteristic curves for the model with the first 2 months of M‐protein dynamic data were ~ 0.8 and 0.85 for POLLUX and CASTOR, respectively. Early M‐protein data within the first 2 months can provide a prospective and reasonable prediction of future long‐term clinical benefit for patients with MM. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
8.
9.
Adrian G. Murphy Marianna Zahurak Mirat Shah Colin D. Weekes Aaron Hansen Lillian L. Siu Anna Spreafico Noelle LoConte Nicole M. Anders Tearra Miles Michelle A. Rudek L. Austin Doyle Barry Nelkin Anirban Maitra Nilofer S. Azad for the ETCTN Study Team 《CTS Clinical and Translational Science》2020,13(6):1178
10.
Hyun A. Lee Seol Ju Moon Hyounggyoon Yoo Mi Kyung Kim Seong Bok Jang Seoungoh Lee Sohee Kim Howard Lee 《CTS Clinical and Translational Science》2021,14(2):625
Gastrointestinal (GI) motility disorders are common, decreases quality of life, and imposes a substantial economic burden. YH12852 is a novel agonist of 5‐hydroxytryptamine for the treatment of GI motility disorders. This phase I/IIa study assessed the tolerability, pharmacodynamic (PD) and pharmacokinetic (PK) profiles of YH12852. In the multiple dose (MD) cohort, healthy subjects and patients with functional constipation were randomized and received orally YH12852 at 0.3, 0.5, 1, 2, or 3 mg or prucalopride 2 mg or their matching placebo, once daily for 14 days after breakfast. In the multiple low‐dose cohort (MLD), healthy subjects randomly received once‐daily oral doses of YH12852 at 0.05 or 0.1 mg for 14 days after breakfast. Questionnaires, gastric emptying breath test for PDs, and plasma samples for PKs were collected. In the MD cohort, a total of 56 subjects (29 healthy volunteers and 27 patients with functional constipation) were randomized, of whom 48 completed the study. In the MLD cohort, a total of 16 healthy subjects were randomized, and 15 subjects completed the study. YH12852 increased the average weekly frequency of spontaneous bowel movements and loosened the stool. In addition, YH12852 increased quality of life satisfaction, and decreased severity of constipation symptom and GI symptoms. YH12852 was safe and well‐tolerated up to 3 mg and showed nearly dose proportional PKs. In conclusion, YH12852 was safe and enhanced GI motility. YH12852 can be developed as an effective treatment option for GI motility disorders, including functional constipation. Further studies are warranted to confirm this possibility. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
11.
Mayur Sarangdhar Mary B. Yacyshyn Andrew R. Gruenzel Melinda A. Engevik Nathaniel L. Harris Bruce J. Aronow Bruce R. Yacyshyn 《CTS Clinical and Translational Science》2021,14(2):518
Recurrent and acute bleeding from intestinal tract angioectasia (AEC) presents a major challenge for clinical intervention. Current treatments are empiric, with frequent poor clinical outcomes. Improvements in understanding the pathophysiology of these lesions will help guide treatment. Using data from the US Food and Drug Administration (FDA)’s Adverse Event Reporting System (FAERS), we analyzed 12 million patient reports to identify drugs inversely correlated with gastrointestinal bleeding and potentially limiting AEC severity. FAERS analysis revealed that drugs used in patients with diabetes and those targeting PPARγ‐related mechanisms were associated with decreased AEC phenotypes (P < 0.0001). Electronic health records (EHRs) at University of Cincinnati Hospital were analyzed to validate FAERS analysis. EHR data showed a 5.6% decrease in risk of AEC and associated phenotypes in patients on PPARγ agonists. Murine knockout models of AEC phenotypes were used to construct a gene‐regulatory network of candidate drug targets and pathways, which revealed that wound healing, vasculature development and regulation of oxidative stress were impacted in AEC pathophysiology. Human colonic tissue was examined for expression differences across key pathway proteins, PPARγ, HIF1α, VEGF, and TGFβ1. In vitro analysis of human AEC tissues showed lower expression of PPARγ and TGFβ1 compared with controls (0.55 ± 0.07 and 0.49 ± 0.05). National Center for Biotechnology Information (NCBI) Gene Expression Omnibus (GEO) RNA‐Seq data was analyzed to substantiate human tissue findings. This integrative discovery approach showing altered expression of key genes involved in oxidative stress and injury repair mechanisms presents novel insight into AEC etiology, which will improve targeted mechanistic studies and more optimal medical therapy for AEC. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
12.
Shufan Ge Susan R. Mendley Jacqueline G. Gerhart Chiara Melloni Christoph P. Hornik Janice E. Sullivan Andrew Atz Paula Delmore Adriana Tremoulet Barrie Harper Elizabeth Payne Susan Lin Jinson Erinjeri Michael CohenWolkowiez Daniel Gonzalez Best Pharmaceuticals for Children Act Pediatric Trials Network Steering Committee 《CTS Clinical and Translational Science》2020,13(6):1189
Metoclopramide is commonly used for gastroesophageal reflux. The aims of the present study were to develop a pediatric population pharmacokinetic (PopPK) model, which was applied to simulate the metoclopramide exposure following dosing used in clinical practice. Opportunistic pharmacokinetic data were collected from pediatric patients receiving enteral or parenteral metoclopramide per standard of care and these data were simultaneously fitted using NONMEM. Allometric scaling with body weight was included a priori in the model. Using the final model, the steady‐state maximum concentrations (Css,max) and the area under the metoclopramide plasma concentration‐time curve at steady state from 0 to 6 hours (AUCss,0–6h) were simulated following 0.1 or 0.15 mg/kg orally every 6 hours in virtual patients, and compared with previously reported ranges associated with toxicity or the efficacy for gastroesophageal reflux in infants. A two‐compartment model with first‐order absorption best characterized 87 concentration measurements from 50 patients (median [range] postnatal age of 8.89 years [0.01–19.13]). There were 20 infants (≤ 2 years), 9 children (2 years to age ≤ 12 years), and 21 adolescents (> 12 years). Body weight was the only covariate included in the final model. For > 75% of virtual patients, simulated Css,max and AUCss,0–6h estimates were within the range associated with efficacy for gastroesophageal reflux in infants; however, slightly lower exposures were predicted in virtual patients < 2 years. Our study suggests that a metoclopramide enteral dose of 0.1 mg/kg every 6 hours, which was previously recommended for pediatric patients, results in simulated exposure generally within suggested ranges for the treatment of gastroesophageal reflux. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
13.
Juhee Kang Jae Whan Kim Hansol Heo Jihyun Lee Kwan Yong Park Jung Han Yoon Jaerak Chang 《CTS Clinical and Translational Science》2021,14(2):606
The current diagnosis of Parkinson’s disease (PD) mostly relies on clinical rating scales related to motor dysfunction. Given that clinical symptoms of PD appear after significant neuronal cell death in the brain, it is required to identify accessible, objective, and quantifiable biomarkers for early diagnosis of PD. In this study, a total of 20 patients with idiopathic PD and 20 age‐matched patients with essential tremor according to the UK Brain Bank Criteria were consecutively enrolled to identify peripheral blood biomarkers for PD. Clinical data were obtained by clinical survey and assessment. Using albumin‐depleted and immunoglobulin G‐depleted plasma samples, we performed immunoblot analysis of seven autophagy‐related proteins and compared the levels of proteins to those of the control group. We also analyzed the correlation between the levels of candidate proteins and clinical characteristics. Finally, we validated our biomarker models using receiver operating characteristic curve analysis. We found that the levels of BCL2‐associated athanogene 2 (BAG2) and cathepsin D were significantly decreased in plasma of patients with PD (P = 0.009 and P = 0.0077, respectively). The level of BAG2 in patients with PD was significantly correlated with Cross‐Culture Smell Identification Test score, which indicates olfactory dysfunction. We found that our biomarker model distinguishes PD with 87.5% diagnostic accuracy (area under the curve (AUC) = 0.875, P < 0.0001). Our result suggests BAG2 and cathepsin D as candidates for early‐diagnosis plasma biomarkers for PD. We provide the possibility of plasma biomarkers related to the autophagy pathway, by which decreased levels of BAG2 and cathepsin D might lead to dysfunction of autophagy. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
14.
Christine Manta Sneha S. Jain Andrea Coravos Dena Mendelsohn Elena S. Izmailova 《CTS Clinical and Translational Science》2020,13(6):1034
The novel coronavirus disease 2019 (COVID‐19) global pandemic has shifted how many patients receive outpatient care. Telehealth and remote monitoring have become more prevalent, and measurements taken in a patient’s home using biometric monitoring technologies (BioMeTs) offer convenient opportunities to collect vital sign data. Healthcare providers may lack prior experience using BioMeTs in remote patient care, and, therefore, may be unfamiliar with the many versions of BioMeTs, novel data collection protocols, and context of the values collected. To make informed patient care decisions based on the biometric data collected remotely, it is important to understand the engineering solutions embedded in the products, data collection protocols, form factors (physical size and shape), data quality considerations, and availability of validation information. This article provides an overview of BioMeTs available for collecting vital signs (temperature, heart rate, blood pressure, oxygen saturation, and respiratory rate) and discusses the strengths and limitations of continuous monitoring. We provide considerations for remote data collection and sources of validation information to guide BioMeT use in the era of COVID‐19 and beyond.In an effort to mitigate the spread of the novel coronavirus disease 2019 (COVID‐19), the disease caused by infection with severe acute respiratory syndrome coronavirus 2 (SARS‐CoV‐2), healthcare providers are increasingly using telehealth for remote patient visits. At the beginning of the pandemic, amidst fears of being infected and having to visit overcrowded hospitals, individuals were rapidly purchasing technologies, such as pulse oximeters, to use at home to monitor for early signs of infection. 1 Entering early summer, the Centers for Disease Control and Prevention (CDC) reported an increase in cases in several regions of the United States; without a vaccine, experts are concerned for a second wave of the virus. 2 , 3 , 4 , 5 As the healthcare system faces an unprecedented need for remote monitoring due to the COVID‐19 pandemic, Biometric Monitoring Technologies (BioMeTs) offer solutions for collecting disease‐related measurements from patients at home. 6 , 7 , 8 BioMeTs are internet‐connected digital medicine products, such as smart thermometers or heart rate monitors with Bluetooth connectivity, that process data captured by mobile sensors using algorithms to generate measures of behavioral and/or physiological function. 9 These connected technologies are used in a variety of contexts, including but not limited to healthcare delivery, 10 clinical trials, 11 and public health. 12 , 13 BioMeTs offer convenient opportunities to collect frequent and objective data and disease‐related measurements, which facilitates assessing trends 12 and detecting changes in vital signs not traceable by conventional spot check data collection protocols. 14 In response to the COVID‐19 pandemic, BioMeTs can be used for many clinical needs, such as aiding preliminary patient physical assessments, assisting with triage of patients with COVID‐19 symptoms, and monitoring patients post‐hospital discharge for risks of readmission. 8 , 15 , 16 , 17 , 18 For clinical teams implementing remote monitoring for the first time or for those already familiar with these technologies and exploring new options, there is an overwhelming variety of BioMeTs available as the market has seen an exponential growth over the past 2 decades. 11 Navigating engineering solutions, form factors (physical size and shape), corresponding data collection protocols, and knowing how to interpret generated values can be challenging, especially if a healthcare provider is unfamiliar with how a BioMeT compares with conventional clinical instruments. Healthcare providers may question the accuracy of measurements taken by patients at home without supervision and it may be unclear how a BioMeT collects and processes data. Understanding data quality and potential biases in data collection is key to drawing appropriate inferences, especially because some of the data may be used for clinical decision making.In this paper, we will discuss the following: (i) sources of information one can use to identify high‐quality BioMeTs, (ii) products and engineering solutions for remote vital sign monitoring, including temperature, heart rate, blood pressure (BP), oxygen saturation, and respiratory rate, and (iii) considerations for choosing a product, including form factors, usability and data collection protocols, and interfering factors that can produce altered readings. Although certain vital sign abnormalities have been associated with COVID‐19 and will be highlighted in this review, we believe the foundations of evaluating these BioMeTs can be applied broadly whenever remote vital sign monitoring is needed. Although overviews of wearable sensor applications for COVID‐19 have been published, 8 , 19 this paper provides a critical review of technologies and is intended as an aid to navigate the plethora of remote monitoring sensors. 相似文献
15.
16.
Katarina Ilic Ivy Song Jingyang Wu Patrick Martin 《CTS Clinical and Translational Science》2020,13(6):1260
Maribavir is an orally bioavailable benzimidazole riboside in clinical development for treatment of cytomegalovirus infection in patients who undergo transplantation. Maribavir was evaluated in a thorough QT (TQT) study to determine any effects on cardiac repolarization. The effect of maribavir 100 and 1,200 mg oral doses on the baseline‐adjusted and placebo‐adjusted corrected QT (QTc) interval (delta delta QTc (ddQTc)) and other electrocardiogram (ECG) parameters was assessed in a randomized, phase I, placebo‐controlled, four‐period crossover study in healthy participants (men and women ages 18–50 years). Additionally, maribavir pharmacokinetics, safety, and tolerability were investigated. Moxifloxacin (400 mg) was used as a positive control to demonstrate the study’s ability to detect QT prolongation. Digital 12‐lead Holter ECG monitoring was performed over 22 hours following study drug administration. Individual, Fridericia’s, and Bazett’s QTc intervals were calculated. Of 52 randomized participants (29 ± 8.1 years old; 31 men (60%)), 50 (96%) completed the study. For both 100‐mg and 1200‐mg doses of maribavir, analysis of ddQTc demonstrated that the upper bound of the two‐sided 90% confidence interval was below the 10‐ms threshold at all time points. The concentration–effect analysis demonstrated no relationship between ddQTc and plasma concentrations of maribavir (and its metabolite). There were no clinically meaningful changes in heart rate and systolic blood pressure. The most common adverse event was dysgeusia; no serious adverse events were reported. This TQT study demonstrated that maribavir did not have impact on cardiac repolarization. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
17.
K. Chris Min Walter K. Kraft Phung Bondiskey Francheska ColnGonzlez Wen Liu Jialin Xu Deborah Panebianco Lori Mixson Marissa F. Dockendorf Catherine Z. Matthews Ramesh Boinpally 《CTS Clinical and Translational Science》2021,14(2):599
Atogepant is a potent, selective, oral calcitonin gene–related peptide (CGRP) receptor antagonist in development for migraine prevention. The chemical structure of atogepant is distinct from previous CGRP receptor antagonists, which were associated with elevated serum alanine aminotransferase (ALT) in clinical trials. Here, we report the safety, tolerability, and pharmacokinetics (PKs) of a once‐daily supratherapeutic dose (170 mg) of atogepant for 28 days from a randomized, double‐blind, placebo‐controlled phase I trial in healthy participants. Overall safety, hepatic safety, and plasma PK parameters were evaluated. Thirty‐four participants aged 23–55 years enrolled; 28 (82.4%) completed the study in accordance with the protocol. Multiple doses of 170 mg atogepant for 28 consecutive days were generally well‐tolerated. All adverse events (AEs; reported in 87.0% of the atogepant group; 72.7%, placebo) were mild in severity except one serious AE of subarachnoid hemorrhage due to a bicycle accident and not considered related to treatment. There were two discontinuations due to AEs, both with atogepant, one considered possibly related to treatment. Over 28 days of treatment, no participant receiving atogepant had an ALT elevation above 1.5 × upper limit of normal. Change from baseline in serum ALT levels was not different between atogepant and placebo. Atogepant is rapidly absorbed (median time to maximum plasma concentration, ~ 2 hours) with an apparent terminal half‐life of ~ 11 hours, and no evidence of accumulation after once‐daily dosing. Overall, atogepant at a high oral dose is safe and well‐tolerated in healthy participants with no clinically meaningful elevations in ALT. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
18.
Karla ClaudioCampos Adaixa Padrn Gabriel Jerkins Jaison Nainaparampil Robyn Nelson Anna Martin Kristin Wiisanen D. Max Smith Yulia Strekalova Michael Marsiske Emily J. Cicali Larisa H. Cavallari Carol A. Mathews 《CTS Clinical and Translational Science》2021,14(2):589
Pharmacogenetic (PGx) testing is a tool to identify patients at a higher risk of adverse events or treatment failure. The concern for unwanted side effects can limit medication adherence, particularly in children and adolescents. We conducted a pragmatic study to evaluate the acceptability and feasibility and gather pilot data on the utility of PGx testing in a child and adolescent psychiatry clinic. Both physicians and families participated in the study and answered pre‐survey and post‐survey questionnaires to examine their attitudes toward PGx testing. Patients were randomized into implementation (N = 25) and control groups (N = 24) and underwent PGx testing at the beginning or end of the study, respectively. Clinical consult notes with genotype‐guided recommendations were provided to physicians for their consideration in clinical decisions. Patient‐reported symptom severity and antidepressant‐related side effects were assessed at baseline and for 12 weeks. Both participating physicians and families agreed that PGx testing is a useful tool to improve medication selection. The time from sample collection to having PGx test results was ~ 10 days and 15 days to having consult notes available, which may have impaired test utility in clinical decision making. There were no differences in any clinical end point between the implementation and control arms; however, there were higher antidepressant side effect scores for CYP2D6 poor and intermediate metabolizers after the eighth week of treatment. Our findings revealed benefits and pitfalls with the use of PGx testing in the real‐world clinical setting, which may inform the methodology of a larger trial focused on outcomes. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?
19.
Christine Keipert Mirco MüllerOlling Franca Gauly Cornelia ArrasReiter Anneliese Hilger 《CTS Clinical and Translational Science》2020,13(6):1127
Emerging treatment options for hemophilia, including gene therapy, modified factor products, antibody‐based products, and other nonreplacement therapies, are in development or on their way to marketing authorization. For proof of efficacy, annual bleeding rates (ABRs) have become an increasingly important endpoint in hemophilia trials. We hypothesized that ABR analyses differ substantially between and within medicinal product classes and that the ABR observation period constitutes a major bias. For ABR characterization, an internal factor VIII (FVIII) treatment database has been built based on confidential clinical trial data submitted to the Paul‐Ehrlich‐Institut (PEI). Furthermore, anonymized data from 46 trial protocols submitted for review to the PEI were analyzed (FVIII replacement, n = 27; antibody‐based, n = 12; and gene therapy, n = 7) for methodology. Definitions of bleeding episodes and ABR observational periods differed substantially in clinical trials. In the initial observation phase, individual ABRs of patients, treated prophylactically for 1 year, vary by about 40% (P < 0.001), which finally led to a significant reduction of the ABR group mean by 20% (P < 0.05). Furthermore, the high variance in ABRs constitutes a major challenge in statistical analyses. In conclusion, considerable heterogeneity and bias in the ABR estimation in clinical trials was identified, which makes it substantially more difficult to compare the efficacy of different treatment regimens and products. Thus, awareness of the important pitfalls when using ABR as a clinical outcome is needed in the evaluation of hemophilia therapies for patients, physicians, regulators, and health technology assessment agencies.Hemophilia is an X‐linked rare bleeding disorder that is characterized by a deficiency of functional coagulation factor VIII (FVIII) or IX and can be categorized based on endogenous factor activity levels as severe (< 1% activity), moderate (1–5% activity), and mild (> 5–40% activity). Individuals with severe hemophilia experience frequent bleeding episodes (BEs) either spontaneously or following minor trauma, which can be acutely life‐threatening or lead to debilitating long‐term complications. For example, joint, muscle, mucosal, and gastrointestinal tract bleeding, and most severely, intracranial hemorrhage can result in disability and death. Current treatment of severe hemophilia mainly relies on replacement therapy with plasma‐derived or modified recombinant factor concentrates.New hemophilia treatment options are in development or have been approved recently, including gene therapy, bispecific monoclonal antibodies, anti‐tissue factor pathway inhibitor antibodies, and other nonreplacement therapies. 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 This is reflected by a large number of ongoing clinical trials (CTs) in this field. In fact, a search in the ClinicalTrials.gov database in June 2019 of phase I–III‐declared studies in congenital hemophilia yielded a total of 69 CTs comprising factor‐based (n = 26), gene therapy‐based (n = 23, including one trial referring to genome editing), antibody‐based (n = 12), RNAi‐based (n = 6), and stem cell‐based (n = 2) products. Importantly, these approaches intervene in different parts of the coagulation cascade and solely coagulation factor levels do not necessarily reflect therapeutic efficacy.Estimation of the annualized bleeding rate, also referred to as annual bleeding rate (ABR), has been introduced early as an efficacy variable for prophylactic replacement therapies in order to complement measures of FVIII or FIX trough levels. However, in contemporary CTs, ABRs are increasingly used as comparative and main outcome parameters.Estimation of bleeding rates has intricate challenges and depends on numerous patient‐related and external factors, including individual clotting factor level, pharmacokinetic profile and pain perception, the subject’s age, health status, activity level, dosing regimen, BE definition, time to follow‐up, and number of patients analyzed. ABR estimation is prone to subjective assessment, as patients as well as treating physicians have to define each bleed. This issue was also demonstrated in a musculoskeletal ultrasound study, which showed that pain perception as well as swelling and warmth is unreliable for bleed detection, resulting in substantial false‐positive and false‐negative bleeding rates. 9 Typically, mean total ABRs are in the low to mid‐single‐digit range, whereas specific ABRs, such as the annual joint bleed rate, are in the low single‐digit range. 10 It has been demonstrated that there is a substantial range of bleeding frequencies among patients with similar clotting factor levels, confirming the ABR as a more personalized parameter. In addition, there is ongoing discussion about the optimal outcome measure and suitability of ABR as an efficacy measure in patients with hemophilia with and without inhibitors. 11 , 12 , 13 , 14 , 15 In the European Medicine Agency (EMA) guidelines on core summary of product characteristics for human plasma derived and recombinant coagulation factor FVIII and FIX products, it is stated that ABR is not comparable between different factor concentrates and between different clinical studies. 16 , 17 This statement has been introduced empirically based on the long‐standing experience in the regulation of hemophilia therapeutics, however, there is lack of supportive and published evidence.We hypothesize that ABR analyses in CTs differ substantially and that the ABR observation period constitutes a major bias. For this approach, we constituted an internal database of confidential FVIII CT data at the Paul‐Ehrlich‐Institut (PEI) to determine basic characteristics of the ABR endpoint. In addition, we analyzed study protocols from contemporary hemophilia CTs comprising replacement and nonreplacement products as well as gene therapies to characterize differences in the methodology of ABR estimation. The results of this study should facilitate guidance on the minimum standards for bleeding rate estimation in CTs of rare bleeding disorders. 相似文献
20.
Jun Chen Richard Perez Angelo Mario de Mattos Cecilia Wang Zhongmin Li Richard L. Applegate II Hong Liu 《CTS Clinical and Translational Science》2020,13(6):1279
Graft function is crucial for successful kidney transplantation. Many factors may affect graft function or cause delayed graft function (DGF), which decreases the prognosis for graft survival. This study was designed to evaluate whether the perioperative use of dexmedetomidine (Dex) could improve the incidence of function of graft kidney and complications after kidney transplantation. A total of 780 patients underwent kidney transplantations, 315 received intravenous Dex infusion during surgery, and 465 did not. Data were adjusted with propensity scores and multivariate logistic regression was used. The primary outcomes are major adverse complications, including DGF and acute rejection in the early post‐transplantation phase. The secondary outcomes included length of hospital stay (LOS), infection, overall complication, graft functional status, post‐transplantation serum creatinine values, and estimated glomerular filtration rate (eGFR). Dex use significantly decreased DGF (19.37% vs. 23.66%; adjusted odds ratio, 0.744; 95% confidence interval, 0.564–0.981; P = 0.036), risk of infection, risk of acute rejection in the early post‐transplantation phase, the risk of overall complications, and LOS. However, there were no statistical differences in 90‐day graft functional status or 7‐day, 30‐day, and 90‐day eGFR. Perioperative Dex use reduced incidence of DGF, risk of infection, risk of acute rejection, overall complications, and LOS in patients who underwent kidney transplantation. Study Highlights
- WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC?
- WHAT QUESTION DID THIS STUDY ADDRESS?
- WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE?
- HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE?