• IMA sites
  • IMAJ services
  • IMA journals
  • Follow us
  • Alternate Text Alternate Text
עמוד בית
Fri, 05.12.25

Search results


March 2023
Ofira Zloto MD, Irit Barequet MD, Orit Ezra Nimni MD, Yoav Berger MD, Juliana Gildener-Leapman MD, Gal Antman MD, Noa Avni-Zauberman MD

Background: The cornea is one of the most densely innervated in the body. Pterygium surgery includes removal of the pterygium tissue from the cornea and conjunctiva followed by autologous conjunctival grafting.

Objectives: To examine the change in corneal and conjunctival sensation post-pterygium surgery.

Methods: This prospective study included patients with primary pterygium. We collected and analyzed demographic data, visual acuity (VA), refraction, quantified sensation, and corneal tomography. Comparison in sensation in the cornea, conjunctiva, and conjunctival autograft was recorded the day of surgery and at least 6 months postoperatively.

Results: Nine patients participated in the study. Mean follow-up time was 9 months (9 ± 3.3, 6–12.4). No complications were documented during or following surgery and no recurrences were found. Statistically significant increases in corneal sensation in the nasal corneal and in the nasal conjunctival areas were noted by the end of follow-up compared to before surgery (P = 0.05, paired samples t-test). There was a significant correlation between the increase in nasal corneal and conjunctival sensation with improved Schirmer testing outcomes and tear break-up time after surgery (P = 0.05, P = 0.01, Pearson correlation). There was a positive correlation between the changes in nasal corneal sensation after surgery and improved changes in VA (P = 0.02, Pearson correlation).

Conclusions: We found improvement in sensation 9 months after pterygium surgery, which may be due to reinnervation of the cornea and conjunctival autograft from the neighboring non-injured nerve fibers. Larger studies with confocal microscopy should be conducted for further analysis.

December 2021
Yana Davidov MD, Yeruham Kleinbaum MD, Yael Inbar MD, Oranit Cohen-Ezra MD, Ella Veitsman MD, Peretz Weiss MD, Mariya Likhter MD, Tania Berdichevski MD PhD, Sima Katsherginsky BA, Avishag Hassid MA, Keren Tsaraf MA, Dana Silverberg BSc, and Ziv Ben Ari MD

Background: New direct acting antiviral agent (DAA) therapies are associated with a high sustained virological response rate (SVR) in hepatitis C virus (HCV) patients. The understanding of the impact of SVR on fibrosis stage is limited.

Objectives: To determine the effect of treatment with the DAAs on long-term liver fibrosis stages, as determined by shear-wave elastography (SWE) or FibroTest©.

Methods: Fibrosis stage was determined at baseline and at 6-month intervals after end of treatment (EOT), using two‐dimensional SWE or FibroTest©; APRI and FIB-4 scores.

Results: The study comprised 133 SVR12 patients. After a median follow-up of 15 months (range 6–33), liver fibrosis stage decreased by at least 1 stage in 75/133 patients (56%). Cirrhosis reversal was observed in 24/82 (29%). Repeated median liver stiffness SWE values in cirrhotic patients were 15.1 kPa at baseline (range 10.5–100), 13.4 kPa (range 5.5–51) at 6 months, and 11.4 kPa (range 6.1–35.8) at 12 months after EOT, P = 0.01. During the second year after EOT, no statistically significant differences in liver fibrosis stage in 12, 18, and 24 months were found. Splenomegaly was the only significant negative predictor of liver fibrosis regression during all time points of repetitive noninvasive assessment.

Conclusions: Following successful DAA treatment, the majority of our HCV patients with advanced fibrosis demonstrated significant improvement, as assessed by non-invasive methods. Advanced fibrosis stage was a negative predictor of fibrosis regression. Longer follow-up periods are required to further establish the impact of DAAs treatment in HCV patients with advanced fibrosis

Noa Avni-Zauberman MD, Barequet S Avni-Zauberma MD, Alon Weissman MD, Juliana Gildener-Leapman MD, Orit Ezra Nimni MD, Yoav Berger MD, and Ofira Zloto MD

Background: Keratoconus is a non-inflammatory disease characterized by progressive corneal steepening, which leads to decreased visual acuity secondary to high irregular astigmatism.

Objectives: To compare the one-year outcomes of accelerated vs. standard collagen crosslinking (CXL) in the treatment of keratoconus.

Methods: A database search of patients who underwent CXL from 2009 to 2017 was conducted at the cornea clinic at Sheba Medical Center. Charts of 99 adult patients (124 eyes) were reviewed. All patients were diagnosed with keratoconus. Main outcome measures were change in keratometry, uncorrected visual acuity (UCVA), and best-corrected visual acuity (BCVA).

Results: We evaluated outcomes in two groups: CXL with standard (3 mW/cm2 for 30 minutes) vs. the accelerated (9 mW/cm2 for 10 minutes) protocol. There were no significant differences between the groups with regard to BCVA, UCVA, and mean spherical equivalent (P =0.83, 0.0519, 0.181, respectively). The corneal thickness in the center and thinnest location were higher in the accelerated group than the in the standard group (P = 0.126). Complication rates did not differ between the two groups.

Conclusions: Accelerated and standard CXL are both safe and effective techniques. Accelerated CXL confers the added benefit of being a faster procedure to both patients and surgeons.

April 2017
Eyal Lotan MD MSc, Stephen P. Raskin MD, Michal M. Amitai MD, Yeruham Kleinbaum MD, Ella Veitsman MD, Peretz Weiss MD, Oranit Cohen-Ezra MD, Tania Berdichevski MD and Ziv Ben-Ari MD

Background: Accurate assessment of liver fibrosis is crucial for the management of patients with hepatitis C virus (HCV) infection.

Objectives: To evaluate the performance of liver segment-to-spleen volume ratio in predicting the severity of liver fibrosis.

Methods: Sixty-four consecutive HCV patients were enrolled in this retrospective study. All patients underwent contrast-enhanced computed tomography (CT) and were divided into three groups based on their hepatic fibrosis stage evaluated by shear-wave elastography (SWE): non-advanced (F0–F1, n=29), advanced (F2, n=19) and severe fibrosis (F3–F4, n=16). Using semi-automated liver segmentation software, we calculated the following liver segments and spleen volumes for each participant: total liver volume (TLV), caudate lobe (CV), left lateral segment (LLV), left medial segment (LMV), right lobe (RV) and spleen (SV), a well as their ratios: CV/SV, RV/SV, LLV/SV, LMV/SV and TLV/SV.

Results: RV/SV was found to discriminate between patients with non-advanced and advanced fibrosis (P = 0.001), whereas SV, CV, RV, TLV/SV, LMV/SV and RV/SV discriminated between patients with advanced and severe fibrosis (P < 0.05). RV/SV ≤ 3.6 and RV ≤ 2.9 were identified as the best cutoff values to differentiate non-advanced from advanced fibrosis and advanced from severe fibrosis with sensitivities of 72.2% and 92.7%, specificities of 72.7% and 77.8%, and with an area under the receiver operating characteristic (ROC) curve of 0.797 and 0.847, respectively (P ≤ 0.002).

Conclusions: RV/SV may be used for the assessment and monitoring of liver fibrosis in HCV patients prior to the administration of antiviral therapy, considering SWE as the reference method.

 

October 2010
H. Duskin-Bitan, S. Kivity, D. Olchovsky, G. Schiby, D. Ezra and M. Mouallem

Background: Kikuchi-Fujimoto disease is a benign and self-limited disease, first reported in Japan in 1972. The characteristic features of this disorder include lymphadenopathy and fever.

Objectives: To summarize our experience with Kikuchi disease with regard to clinical manifestations and outcome.

Methods: The patients included in the study were those diagnosed with Kikuchi disease during the years 2005–2008 in two departments of internal medicine at Sheba Medical Center.

Results: We identified five patients with Kikuchi disease; four of them were women and the mean age was 22.6 years. All the patients had cervical lymphadenopathy; three had other sites of lymphadenopathy. Four of the patients had fever higher than 39ºC. Two of them had splenomegaly and three reported weight loss. Three of the five patients experienced a relapse of the disease and were treated with steroids or non-steroidal anti-inflammatory agents. The diagnosis was confirmed in all the patients by an excisional biopsy of lymph node.

Conclusions: Kikuchi disease must be considered in every young patient with fever and lymphadenopathy. The disease usually has a benign course.

September 2010
G. Twig, A. Lahad, I. Kochba, V. Ezra, D. Mandel, A. Shina, Y. Kreiss and E. Zimlichman

Background: A survey conducted among Israel Defense Force primary care physicians in 2001 revealed that they consider patients' needs more than they do organizational needs and that the education PCPs[1] currently receive is inadequate. In 2003 the medical corps initiated a multi-format continuous medical education program aimed at improving skills in primary care medicine.

Objectives: To measure and analyze the effect of the tailored-made CME[2] program on PCPs’ self-perception 3 years after its implementation and correlate it to clinical performance.

Methods: In 2006 a questionnaire was delivered to a representative sample of PCPs in the IDF[3]. The questionnaire included items on demographic and professional background, statements on self-perception issues, and ranking of roles. We compared the follow-up survey (2006) to the results of the original study (2001) and correlated the survey results with clinical performance as measured through objective indicators.

Results: In the 2006 follow-up survey PCPs scored higher on questions dealing with their perception of themselves as case managers (3.8 compared to 4.0 on the 2001 survey on a 5 point scale, P = 0.046), perceived quality of care and education (3.5 vs. 3.8, P = 0.06), and on questions dealing with organizational commitment (3.5 vs. 3.8, P=0.01). PCPs received higher scores on clinical indicators in the later study (odds ratio 2.05, P < 0.001).

Conclusions: PCPs in the IDF perceive themselves more as case managers as compared to the 2001 survey. A tailor-made CME program may have contributed to the improvement in skills and quality of care.






[1] PCP = primary care physician



[2] CME = continuous medical education



[3] IDF = Israel Defense Forces


December 2009
S. Weitzman, S. Greenfield, J. Billimek, H. Tabenkin, P. Schvartzman, E. Yehiel, H. Tandeter, S. Eilat‎-Tsanani and S.H. Kaplan

Background: Research on synergistic effects of patient-targeted interventions combined with physician-targeted interventions has been limited.

Objectives: To compare a combined physician-patient intervention to physician feedback alone on a composite outcome of glycemic, lipid and blood pressure control.

Methods: In this cluster study 417 patients with adult-type 2 diabetes from four primary care clinics were randomized to receive either a physician-only intervention or a combined physician-plus-patient intervention. Physicians in all clinics received diabetes-related quality performance feedback during staff meetings. Patients at combined-intervention clinics also received a letter encouraging them to remind their doctors to address essential aspects of diabetes care at the next visit. At 1 year follow-up, outcome measurements included hemoglobin A1c, low density lipoprotein-cholesterol and systolic blood pressure; the proportion of patients with HbA1c[1] < 9%, LDL[2] < 130 mg/dl and SBP[3] < 140 mmHg both as separate outcomes and combined.

Results: After adjusting for patient characteristics and baseline measures, follow-up levels of HbA1c (7.5% vs. 7.8%, P = 0.09), LDL (104.7 vs. 110.7 mg/dl, P < 0.05) and SBP (135.6 vs. 139.9, P = 0.10) were marginally better for combined-intervention patients compared to physician-only intervention patients. Significantly more patients in the combined-intervention (38.8%) than physician-only intervention (24.2%) met all three targets (HbA1c < 9%, LDL < 130 mg/dl and SBP < 140 mmHg) as a single combined outcome (adjusted odds ratio 2.4, P < .01).

Conclusions: Compared to physician-feedback alone, a dual intervention combining a patient letter with physician feedback produced modest improvements in glycemic, lipid and blood pressure control individually, but substantial improvement in a combined measure of these three outcomes together. Using composite outcomes may detect meaningful improvements in the management of complex chronic disease. 


 




[1] HbA1c = hemoglobin A1c



[2] LDL = low density lipoprotein



[3] SBP = systolic blood pressure


November 2009
U. Nussinovitch, D. Ezra, N. Nussinovitch and Y. Shoenfeld
November 2008
R. Loebstein et al

Background: Infections with blood-borne viruses are a major health problem among illicit drug users. There is little information about infection rates and risk factors for hepatitis virus B, C or the human immunodeficiency virus in drug users in Israel.

Objectives: To determine the prevalence of HCV[1], HBV[2] and HIV[3] infections in a large cohort of drug users in Israel; to compare rates of HCV, HBV and HIV between injecting versus non-injecting drug users and between different origin countries; and to identify risk factors for HCV among illicit drug users.

Methods: We conducted a cross-sectional interviewer-administered questionnaire and serological screening for HCV, HBV and HIV in 1443 consecutive drug users diagnosed at the Israeli National Center for Diagnosis of Drug Addicts between January 2003 and December 2005.

Results: Fourteen (0.9%), 51 (3.5%) and 515 (35.7%) subjects tested positive for HIV, HBV and HCV, respectively. All three infections (HIV, HBV and HCV) were significantly more common among injecting drug users and immigrants from the former Soviet Union and other East European countries compared to native Israelis. Multivariate analysis showed that HCV infection was associated with age (> 40 years) (OR=2.06, 95% CI 1.40–3.03), immigration from East European countries and the former Soviet Union (OR=4.54, 95% CI 3.28–6.28), and injecting drug use (OR=16.44, 95% CI 10.79–25.05).

Conclusions: HIV, HBV and HCV prevalence among drug users in Israel is significantly lower than in North America and West Europe. Risk factors for HCV infection in this population include injecting drug use, older age, and immigration from the former Soviet Union.






[1] HCV = hepatitis C virus

[2] HBV = hepatitis B virus

[3] HIV = human immunodeficiency virus


December 2006
U. Elchalal, E. Gabbay, M. Nadjari, D. Varon, O. Zelig and E. Ben-Chetrit
October 2006
V.H. Eisenberg, D. Raveh, Y. Meislish, B. Rudensky, Y. Ezra, A. Samueloff, A.I. Eidelman and M.S. Schimmel
 Background: Previous assessments of maternal group B Streptococcus carrier rates in women delivering at Shaare Zedek Medical Center ranged between 3.5 and 11% with neonatal sepsis rates of 0.2–0.9/1000 live births. Because of low colonization and disease rates, routine prenatal cultures of GBS[1] were not recommended, and intrapartum prophylaxis was mainly based on maternal risk factors.

Objectives: To determine whether this policy is still applicable. 

Methods: We performed prospective sampling and follow-up of women admitted for labor and delivery between February 2002 and July 2002. Vaginal and rectal cultures were obtained before the first pelvic examination. GBS isolation was performed using selective broth medium, and identified by latex agglutination and serotyping. Demographic data were collected by means of a standardized questionnaire. Data on the newborns were collected throughout 2002.

Results: Of the 629 sampled women, 86 had a positive culture and a carrier rate of 13.7%. A borderline significantly higher carriage rate was observed among mothers of North American origin (21% vs. 13.1%, P = 0.048), and a higher attack rate in their infants (3.8/1000 compared with 0.5/1000 live births in our general maternal population, P = 0.002). Eight newborns had early-onset neonatal GBS sepsis (a rate of 0.8/1000 live births), but none of them benefited from intrapartum antibiotic prophylaxis.

Conclusions: An increased neonatal disease rate was observed in a population with a higher colonization rate than previously seen. In lieu of the higher carrier rates, we now recommend routine prenatal screening for GBS in our perinatal population.


 





[1] GBS = group B Streptococcus


May 2005
E. Segev, D. Keret, F. Lokiec, A. Yavor, S. Wientroub, E. Ezra and S. Hayek
 Background: The preferred conservative treatment for congenital idiopathic clubfoot deformity remains a controversial issue.


Objectives: To compare the outcomes of two groups of CICF[1] patients who underwent two different treatment protocols.

Methods: The study cohort included 72 infants treated in our hospital from August 1998 to December 2002. Group 1 comprised 40 infants (61 clubfeet) who were treated by a traditional method (a modification of the Kite and Lovell technique), and group 2 consisted of 32 infants (48 clubfeet) who were treated with the Ponseti technique. Both groups were similar in age, gender and severity of the deformity (Dimeglio scoring system)

Results: After an average follow-up of 54.9 months (range 44–68), 35 (57%) clubfeet in group 1 required surgical intervention and 27 (44%) clubfeet had a residual deformity at last follow-up. In the Ponseti group, 45 (94%) clubfeet were fully corrected at last follow-up (average 29.2 months, range 16–45) while 3 (6%) clubfeet had residual deformity and required surgery. Tendo-Achilles tenotomy was performed with no complications in 47 clubfeet (in group 2) at an average age of 2.4 months (range 2–4 months).

Conclusions: Even after a relatively short follow‑up period, our success rate with the Ponseti approach already appears to be significantly higher and to bear fewer complications than the traditional treatment, in agreement with the results published by other medical centers. We now endorse the Ponseti technique of conservative manipulative treatment for congenital idiopathic clubfoot in our department.


 





[1] CICF = congenital idiopathic clubfoot


December 2004
I. Solomon, N. Maharshak, G. Chechik, L. Leibovici, A. Lubetsky, H. Halkin, D. Ezra and N. Ash

Background: Oral anticoagulation with warfarin can lead to life-threatening events as a result of either over-anticoagulation or undertreatment. One of the main contributors to an undesirable warfarin effect is the need to adjust its daily dose for a specific patient. The dose is adjusted empirically based on the experience of the clinician, a method that is often imprecise. There is currently no other well-accepted method for predicting the maintenance dose of warfarin.

Objective: To describe the application of an artificial neural network to the problem of warfarin maintenance dose prediction.

Methods: We designed a neural network that predicts the maintenance dose of warfarin. Data on 148 patients attending a large anticoagulant clinic were collected by file review. Using correlational analysis of the patients' data we selected the best input variables. The network was trained by using the back-propagation algorithm on a subset of our data and the results were validated against the rest of the data. We used a multivariate linear regression to create a comparable model.

Results: The neural network generated reasonable predictions of the maintenance dose (r = 0.823). The results of the linear regression model were similar (r = 0.800).

Conclusion: Neural networks can be applied successfully for warfarin maintenance dose prediction. The results are promising, but further investigation is needed.
 

E. Segev, E. Ezra, Y. Binyamini, S. Weintroub and J. Ben-Chaim

Background: Bladder exstrophy is a severe congenital defect that requires a multidisciplinary treatment approach. Soft tissue repair may be successful during the first few days after birth but a combination of pelvic osteotomies and bladder reconstruction is necessary later in life. The combination of externally fixed anterior and posterior osteotomies has biomechanical advantages over previous techniques for achieving primary bladder closure.

Objectives: To describe our experience with a combined vertical and horizontal pelvic osteotomy approach for the repair of bladder exstrophy.

Methods: Four children underwent bladder exstrophy closure; the mean age at surgery was 19 months (range 9–33 months)... We stabilized the osteotomies with a small Synthes AO external fixator, 4.0 mm rod diameter.

Results: All four patients had successful bladder repair with no dehiscence; two of them achieved partial continence, and bladder neck reconstruction is planned for the other two. Three of the four patients sustained neurologic injury; two completely recovered, and the third continues to suffer from right drop foot. The average follow-up was 39 months (range 10–60 months).


Conclusions: Vertical and horizontal pelvic osteotomies stabilized by external fixator and bladder repair is an effective treatment for bladder exstrophy.

September 2003
A.B. Chkhotua, T. Klein, E. Shabtai, A. Yussim, N. Bar-Nathan, E. Shaharabani, S. Lustig and E. Mor

Background: Recent advances in immunosuppressive therapy have led to a substantial improvement in the outcome of kidney transplantation. Living unrelated donors may become a source of additional organs for patients on the kidney waiting list.

Objectives: To study the impact of combination of calcineurin inhibitors and mycophenolate-mofetile, together with steroids, on outcomes of living related and unrelated transplants. 

Methods: Between September 1997 and January 2000, 129 patients underwent living related (n=80) or unrelated (n=49) kidney transplant. The mean follow-up was 28.2 months. Immunosuppressive protocols consisted of MMF[1] with cyclosporine (41%) or tacrolimus (59%), plus steroids. Patient and graft survival data, rejection rate, and graft functional parameters were compared between the groups.

Results: LUD[2] recipients were older (47.8 vs. 33.6 years) with higher number of re-transplants (24.5% vs. 11.2% in LRD[3] recipients, P < 0.05). Human leukocyte antigen matching was higher in LRD recipients (P < 0.001). Acute rejection developed in 28.6% of LUD and 27.5% of LRD transplants (P = NS). Creatinine levels at 1, 2 and 3 years post-transplant were 1.6, 1.7 and 1.7 mg/dl for LRD patients and 1.5, 1.5 and 1.3 mg/dl for LUD recipients (P = NS). There was no difference in patient survival rates between the groups. One, 2 and 3 year graft survival rates were similar in LRD (91.3%, 90% and 87.5%) and LUD (89.8%, 87.8% and 87.8%) recipients.

Conclusions: Despite HLA[4] disparity, rejection and survival rates of living unrelated transplants under current immunosuppressive protocols are comparable to those of living related transplants.






[1] MMF = mycophenolate-mofetile



[2] LUD = living unrelated donor



[3] LRD = living related donor



[4] HLA = human leukocyte antigen


Legal Disclaimer: The information contained in this website is provided for informational purposes only, and should not be construed as legal or medical advice on any matter.
The IMA is not responsible for and expressly disclaims liability for damages of any kind arising from the use of or reliance on information contained within the site.
© All rights to information on this site are reserved and are the property of the Israeli Medical Association. Privacy policy

2 Twin Towers, 35 Jabotinsky, POB 4292, Ramat Gan 5251108 Israel