HIV is a major public health issue affecting millions globally. Women and girls account for 46% of new HIV infections in 2022 and approximately 1.3 million females become pregnant every year. Vertical transmission of HIV from persons living with HIV (PLHIV) to infants may occur through different modalities, such as through breast/chest feeding. Notably, 82% of PLHIV who chose to breast/chest feed are on antiretroviral therapy (ART) when feeding their infants. Precise estimates of the risk of postpartum transmission to infants during breast/chest feeding at varying viral load levels remain a significant gap in the literature.
A rapid systematic search of electronic databases will be conducted from January 2005 to the present, including Medline, Embase and Global Health. The objective of this rapid review is to explore and assess the available evidence on the effect of varying viral load levels on the risk of HIV transmission to infants during breast/chest feeding when the birthing or gestational parent living with HIV is on ART. Study characteristics will be summarised and reported to support the narrative summary of the findings. The focus will be on the absolute risk of HIV transmission from birthing parent to infant during chest/breast feeding. The findings will also be stratified by month, including the risk of HIV transmission for 6 months and greater than 6 months postpartum. We will ascertain the risk of bias using A Measurement Tool to Assess Systematic Reviews 2, Quality of Prognosis Studies and Downs and Black checklist for the appropriate study type. A summary score will not be calculated, rather the strengths and limitations of the studies will be narratively described.
No human subjects will be involved in the research. The findings of this rapid review will inform a future systematic review and will be disseminated through peer-reviewed publications, presentations and conferences.
CRD42024499393.
The Modern Innovative Solutions to Improve Outcomes in Asthma, Breathlessness and Chronic Obstructive Pulmonary Disease (COPD) (MABC) service aimed to enhance disease management for chronic respiratory conditions through specialist multidisciplinary clinics, predominantly in the community. This study assesses the outcomes of these clinics.
This study used a prospective, longitudinal, participatory action research approach.
The study was conducted in primary care practices across Hampshire, UK.
Adults aged 16 years and above with poorly controlled asthma or COPD, as well as those with undifferentiated breathlessness not under specialist care, were included.
Participants received care through the multidisciplinary, specialist-led MABC clinics.
Primary outcomes included disease activity, quality of life and healthcare utilisation. Secondary outcomes encompassed clinic attendance, diagnostic changes, patient activation, participant and healthcare professional experiences and cost-effectiveness.
A total of 441 participants from 11 general practitioner practices were recruited. Ninety-six per cent of participants would recommend MABC clinics. MABC assessments led to diagnosis changes for 64 (17%) participants with asthma and COPD and treatment adjustments for 252 participants (57%). Exacerbations decreased significantly from 236 to 30 after attending the clinics (p
Specialist-supported multidisciplinary teams in MABC clinics improved diagnosis accuracy and adherence to guidelines. High patient satisfaction, disease control improvements and reduced exacerbations resulted in decreased unscheduled healthcare use and cost savings.
by Yukiko Morisaki, Noriyuki Miyata, Megumi Nakashima, Tomokazu Hata, Shu Takakura, Kazufumi Yoshihara, Takafumi Suematsu, Koji Nomoto, Kouji Miyazaki, Hirokazu Tsuji, Nobuyuki Sudo
Recent evidence suggests a crucial role of the gut microbiota in the pathogenesis of anorexia nervosa (AN). In this study, we carried out a series of multiple analyses of the gut microbiota of hospitalized individuals with AN over three months using 16S or 23S rRNA-targeted reverse transcription–quantitative polymerase chain reaction (PCR) technology (YIF-SCAN®), which is highly sensitive and enables the precise quantification of viable microorganisms. Despite the weight gain and improvements in psychological features observed during treatment, individuals with AN exhibited persistent gut microbial dysbiosis over the three-month duration. Principal component analysis further underscored the distinct microbial profile of individuals with AN, compared with that of age-matched healthy women at all time points. Regarding the kinetics of bacterial detection, the detection rate of Lactiplantibacillus spp. significantly increased after inpatient treatment. Additionally, the elevation in the Bifidobacterium counts during inpatient treatment was significantly correlated with the subsequent body weight gain after one year. Collectively, these findings suggest that gut dysbiosis in individuals with AN may not be easily restored solely through weight gain, highlighting the potential of therapeutic interventions targeting microbiota via dietary modifications or live biotherapeutics.by Douglas C. Smith, Crystal A. Reinhart, Shahana Begum, Janaka Kosgolla, John F. Kelly, Brandon B. Bergman, Marni Basic
Background and aimsTo date, no epidemiological survey has estimated the prevalence of adolescents identifying as being in recovery. This is necessary for planning and identifying the needs of youth with current and remitted substance use disorders. This study estimated the prevalence of recovery status in a large statewide epidemiological survey administered between January and March 2020.
ParticipantsParticipants were high school students in 9th through 12th grades throughout Illinois.
MeasurementsYouth were asked if they were in recovery and if they had resolved problems with substances. Youth who reported recovery and problem resolving dual status (DS), recovery only (RO), and problem resolution only (PRO) were compared to propensity score matched control groups who reported neither status (neither/nor; NN). Outcomes included alcohol use, binge alcohol use, cannabis use, and prescription drug use in the past 30 days.
FindingsPrevalence estimates were 884 (1.4%) for DS, 1546 (2.5%) for PRO, and 1,811 (2.9%) for RO. Relative to propensity matched control samples, all three groups had significantly lower odds of prescription drug use. The PRO group had lower odds of past month cannabis use. There were no significant differences for either alcohol outcome.
ConclusionsPrevalence estimates of youth in recovery are slightly lower than those of adults in recovery, and estimates should be replicated. Youth in recovery and those resolving problems have numerous behavioral health needs, and relative to matched controls, have even odds for past 30-day alcohol use. These findings compel us to further define recovery for adolescents and emerging adults to allow for improving treatments and epidemiological research.
by Delphine Rapp, Colleen Ross, Vanessa Cave, Paul Maclean, Ruy Jauregui, Gale Brightwell
Land-spreading of animal faecal wastes -such as animal beddings- can introduce zoonotic enteropathogens into the food system environment. The study evaluated the effectiveness of animal beddings naturally contaminated by calf manure to reduce E. coli O157:H7 or Salmonella enterica. The two pathogens were introduced separately as a four strains-cocktail and at high (>6.5 Log10 g-1) concentration into bedding materials, and their inactivation over a 10 weeks-period was monitored by using a Most Probable Number (MPN) enumeration method. Inactivation of E. coli O157:H7 was more effective in the bedding inoculated immediately after collection from calf pens than in the beddings inoculated after a 2 months-pre-storage period: E. coli O157:H7 levels were reduced by 6.6 Log10 g-1 in unstored bedding (0.5 Log10 g-1 recovered; 95%CI: 0.0–1.2), and by 4.9 Log10 g-1 in pre-stored bedding (2.2 Log10 g-1 recovered; 95%CI: 1.5–2.8) with a significant (pS. enterica was inactivated less effectively as counts were reduced by one order of magnitude, with no significant difference in inactivation between unstored and pre-stored beddings. Low levels of naturally occurring E. coli O157 and Salmonella spp. were detected in the non-inoculated beddings, as well as in the straw prior to use in the animal facility. To better understand the possible biological processes involved, the bacterial community present in the beddings was characterised by short-read 16S rRNA sequencing. Pre-storage of the bedding affected the composition but not the diversity of the bacterial community. Analyses of the key bacterial phyla suggested that the presence of a diverse and stable bacterial community might facilitate inactivation of the introduced pathogens, and a possible role of bacterial orders associated with lignocellulolytic resources. Overall, the study contributed to the understanding of the fate of zoonotic bacteria introduced in animal beddings during storage and identified bedding storage practices pre-and post-use in animal facilities that could be important to prevent the risk of zoonosis dissemination to the environment or to the dairy herds.HIV-induced chronic inflammation, immune activation and combination antiretroviral therapy (cART) are linked with adverse metabolic changes known to cause cardiovascular adversities. This study evaluates the prevalence of lipodystrophy, and metabolic syndrome (MetS), and analyses risk factors in HIV-infected Ethiopians taking cART.
A multicentre cross-sectional study was conducted at tertiary-level hospitals. Eligible participants attending the HIV clinics were enrolled. Sociodemographic, anthropometric, clinical, HIV treatment variables, lipid profile, fasting blood glucose level, risk factors and components of MetS, also lipodystrophy, were studied. Data were analysed by SPSS statistical package V.25 with descriptive and analytical statistics. For multivariable analysis of risk factors, a logistic regression model was used. Results were presented in frequency and percentages, mean±SD, or median+IQR. Statistical significance was taken as p
Among 518 studied participants, two-thirds were females, and the mean age of the study population was 45 years (SD=11). The mean duration of cART was 10 years (SD=4). Median CD4 count was 460 cells/mm3. The prevalence of MetS according to the Adult Treatment Panel III (2005) criteria was 37.6%. In multivariable analysis, independent risk factors for MetS were age >45 years (aHR 1.8, 95% CI 1.2 to 2.4), female sex (aHR 1.8, 95% CI 1.1 to 2.8), body mass index (BMI)
Our study revealed 38% of the participants had MetS indicating considerable cardiovascular disease (CVD) risks. Independent risk factors for MetS were BMI≥25 kg/m2, efavirenz and lopinavir/ritonavir-based cART, female sex and age ≥45 years. In addition to prevention, CVD risk stratification and management will reduce morbidity and mortality in people with HIV infection.
by John Efromson, Giuliano Ferrero, Aurélien Bègue, Thomas Jedidiah Jenks Doman, Clay Dugo, Andi Barker, Veton Saliu, Paul Reamey, Kanghyun Kim, Mark Harfouche, Jeffrey A. Yoder
Normal development of the immune system is essential for overall health and disease resistance. Bony fish, such as the zebrafish (Danio rerio), possess all the major immune cell lineages as mammals and can be employed to model human host response to immune challenge. Zebrafish neutrophils, for example, are present in the transparent larvae as early as 48 hours post fertilization and have been examined in numerous infection and immunotoxicology reports. One significant advantage of the zebrafish model is the ability to affordably generate high numbers of individual larvae that can be arrayed in multi-well plates for high throughput genetic and chemical exposure screens. However, traditional workflows for imaging individual larvae have been limited to low-throughput studies using traditional microscopes and manual analyses. Using a newly developed, parallelized microscope, the Multi-Camera Array Microscope (MCAM™), we have optimized a rapid, high-resolution algorithmic method to count fluorescently labeled cells in zebrafish larvae in vivo. Using transgenic zebrafish larvae, in which neutrophils express EGFP, we captured 18 gigapixels of images across a full 96-well plate, in 75 seconds, and processed the resulting datastream, counting individual fluorescent neutrophils in all individual larvae in 5 minutes. This automation is facilitated by a machine learning segmentation algorithm that defines the most in-focus view of each larva in each well after which pixel intensity thresholding and blob detection are employed to locate and count fluorescent cells. We validated this method by comparing algorithmic neutrophil counts to manual counts in larvae subjected to changes in neutrophil numbers, demonstrating the utility of this approach for high-throughput genetic and chemical screens where a change in neutrophil number is an endpoint metric. Using the MCAM™ we have been able to, within minutes, acquire both enough data to create an automated algorithm and execute a biological experiment with statistical significance. Finally, we present this open-source software package which allows the user to train and evaluate a custom machine learning segmentation model and use it to localize zebrafish and analyze cell counts within the segmented region of interest. This software can be modified as needed for studies involving other zebrafish cell lineages using different transgenic reporter lines and can also be adapted for studies using other amenable model species.by Carlos Ignacio Reyes-Farias, Marcelo Reategui-Diaz, Franco Romani-Romani, Larry Prokop
BackgroundSodium-glucose cotransporter 2 (SGLT-2) inhibitors have shown a favorable effect on cardiovascular and renal outcomes in patients with type 2 diabetes mellitus (T2DM). However, their efficacy in patients with chronic kidney disease (CKD) with or without T2DM has not yet been analyzed.
ObjectiveTo assess the cardiovascular and renal effects of SGLT-2 inhibitors in patients with CKD with and without T2DM, including all CKD patients in the current literature.
MethodsWe searched MEDLINE, EMBASE, CENTRAL and Scopus for randomized controlled trials of SGLT-2 inhibitors that evaluated cardiovascular and kidney outcomes in patients with CKD, or trials in which these patients were a subgroup. We defined 2 primary outcomes: a composite of cardiovascular death or hospitalization for heart failure, and a composite renal outcome. For each outcome, we obtained overall hazard ratios with 95% confidence intervals by using a random effects model.
ResultsWe included 14 randomized controlled trials. SGLT-2 inhibitors decreased the hazard for the primary cardiovascular outcome (HR 0.76; [95% CI 0.72–0.79]) and the primary renal outcome (HR 0.69; [95% CI 0.61–0.79]) in patients with CKD with or without T2DM. We did not find significant differences in the subgroup analyses according to diabetes status, baseline eGFR values or the type of SGLT-2 inhibitor used.
ConclusionIn patients with CKD, treatment with SGLT-2 inhibitors in addition to standard therapy conferred protection against cardiovascular and renal outcomes. Further research on patients with non-diabetic CKD should be done to confirm the utility of these medications in this population. (PROSPERO ID: CRD42021275012).
Circulatory system disease (CSD) patterns vary over time and between countries, related to lifestyle risk factors, associated in turn with socioeconomic circumstances. Current global CSD epidemics in developing economies are similar in scale to those observed previously in the USA and Australasia. Australia exhibits an important macroeconomic phenomenon as a rapidly transitioning economy with high immigration throughout the nineteenth and twentieth centuries. We wished to examine how that historical immigration related to CSD patterns subsequently.
We provide a novel empirical analysis employing census-derived place of birth by age bracket and sex from 1891 to 1986, in order to map patterns of immigration against CSD mortality rates from 1907 onwards. Age-specific generalised additive models for both CSD mortality in the general population, and all-cause mortality for the foreign-born (FB) only, from 1910 to 1980 were also devised for both males and females.
The percentage of FB fell from 32% in 1891 to 9.8% in 1947. Rates of CSD rose consistently, particularly from the 1940s onwards, peaked in the 1960s, then declined sharply in the 1980s and showed a strong period effect across age groups and genders. The main effects of age and census year and their interaction were highly statistically significant for CSD mortality for males (p
We argue our empirical calculations, supported by historical and socioepidemiological evidence, employing immigration patterns as a proxy for epidemiological transition, affirm the life course hypothesis that both early life circumstances and later life lifestyle drive CSD patterns.
The dynamics of self-harm vary substantially around the world, yet it is severely under-researched outside of a small number of high-income ‘Western’ countries. South Asia is disproportionately impacted by suicide, yet we know less about non-suicidal self-injury (NSSI) in the region.
To review and summarise evidence on the prevalence, correlates and common methods of NSSI in South Asia.
We searched Medline, Embase and PsycINFO for the period 1 January 2000 to 31 August 2023, for peer-reviewed observational studies. A total of 11 studies from eight South Asian countries that reported prevalence and/or correlates of NSSI were included in this review. We assessed the quality of the studies using the Study Quality Assessment Tools for Observational Cohort and Cross-Sectional Studies. We used meta-regression to describe the sources of heterogeneity. Abstracted data were summarised using thematic synthesis.
For non-clinical populations, the 12-month prevalence of NSSI ranged from 3.2% to 44.8%, and the lifetime prevalence ranged from 21% to 33%. For clinical populations, the 12-month prevalence of NSSI ranged from 5% to 16.4%, while the lifetime prevalence ranged from 2% to 27%. Male sex, unemployment, financial stress, history of suicidal behaviour and depression were associated with a higher risk of NSSI. Better access to counselling services, higher self-esteem and self-knowledge were associated with a lower risk of NSSI.
The burden of NSSI in South Asia appears to be high in both clinical and non-clinical populations. Further research, especially with general population samples, is needed to build evidence on the epidemiology, context and meaning of NSSI in South Asia to inform the design of context-specific interventions.
CRD42022342536.
This study aimed to assess the prognostic significance of residual (discharge) dyspnoea in acute heart failure (AHF) patients.
Single-centre, prospective observational study.
Patients hospitalised for decompensated AHF in a single cardiology centre, in Poland.
All patients (n=202) who survived the hospitalisation with the primary diagnosis of AHF and were discharged from the hospital.
1-year all-cause mortality; and the composite endpoint of 1-year all-cause mortality and rehospitalisation for the HF (whichever occurred first).
On admission, 159 (78.7%) AHF patients presented dyspnoea at rest, while residual resting dyspnoea at discharge was present in 16 patients (7.9%). There were 48 (24%) patients with moderate/severe exertional dyspnoea at discharge. In the multivariable model, the resting dyspnoea at discharge was related to a higher risk of both 1-year mortality and composite outcome, with HR (95% CI) 8.0 (3.7 to 17.3) and 5.1 (2.6 to 10.2), respectively, both p
Among AHF patients the residual dyspnoea at discharge was unexpectedly common and was associated with an unfavourable outcome during 1-year follow-up.
by Ana Castro-Avila, Catalina Merino-Osorio, Felipe González-Seguel, Agustín Camus-Molina, Felipe Muñoz-Muñoz, Jaime Leppe, on behalf of the IMPACCT COVID-19 study group
IntroductionThe COVID-19 pandemic can be seen as a natural experiment to test how bed occupancy affects post-intensive care unit (ICU) patient’s functional outcomes. To compare by bed occupancy the frequency of mental, physical, and cognitive impairments in patients admitted to ICU during the COVID-19 pandemic.
MethodsProspective cohort of adults mechanically ventilated >48 hours in 19 ICUs from seven Chilean public and private hospitals. Ninety percent of nationwide beds occupied was the cut-off for low versus high bed occupancy. At ICU discharge, 3- and 6-month follow-up, we assessed disability using the World Health Organization Disability Assessment Schedule 2.0. Quality of life, mental, physical, and cognitive outcomes were also evaluated following the core outcome set for acute respiratory failure.
ResultsWe enrolled 252 participants, 103 (41%) during low and 149 (59%) during high bed occupancy. Patients treated during high occupancy were younger (P50 [P25-P75]: 55 [44–63] vs 61 [51–71]; p Conclusions
There were no differences in post-ICU outcomes between high and low bed occupancy. Most patients (>90%) had at least one mental, physical or cognitive impairment at ICU discharge, which remained high at 6-month follow-up (57%).
Clinical trial registrationNCT04979897 (clinicaltrials.gov).