首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
We evaluated survival and renal function of cadaveric donor grafts according to donor age. The median age of the pediatric donors was 7.0 (0.7-16) years in 46 patients [median age 11.8 years (range) 3-16.8 years]. The median age of the adult donors was 34.4 (19-54) years in 59 patients [median age 12.1 years (range) 7-17.3 years]. Thirty patients were treated with azathioprine and prednisolone and 75 with cyclosporine A and prednisolone. The glomerular filtration rate (GFR) and the effective renal plasma flow (ERPF) were determined by the clearances of 51chromium-EDTA and 125iodine-hippurate 1-48 months after kidney transplantation. There was no difference in graft survival between pediatric and adult grafts. There were also no differences in GFR in patients receiving grafts from pediatric or adult donors; 2-3 months after transplantation the GFR in recipients of pediatric grafts was 62 +/- 20 ml/min per 1.73 m2 compared with 61 +/- 21 in those receiving adult grafts. The ERPF in recipients of adult grafts was significantly higher in the 1st month after transplantation: 486 +/- 239 versus 362 +/- 158 ml/min per 1.73 m2. From the 4th to the 6th month after transplantation this difference disappeared: the ERPF of grafts from pediatric donors was 279 +/- 131 ml/min per 1.73 m2 compared with 273 +/- 123 ml/min per 1.73 m2 in grafts from adult donors. Using the single-kidney GFR and ERPF on an age-matched group of probands with minor diseases as references, 2-3 months after transplant the mean GFR of grafts from pediatric donors increased to 118% +/- 51%, whereas the GFR of adult donor grafts fell to 60% +/- 22% over the same period. After 4-6 months the ERPF in pediatric grafts was 96% +/- 55% compared with 50% +/- 22% in adult grafts. We conclude that graft survival and function in children with either a pediatric or an adult graft may not differ because graft function adapts to the requirement of the recipient.  相似文献   

2.
Renal function was studied in 2 groups of renal transplant recipients and their donors by technetium-99m diethylenetriamine pentaacetic acid and a gamma camera. The pediatric group (group A) comprised 40 children and their adult kidney donors. The adult group (group B) consisted of 112 consecutive adult renal transplant recipients and their adult donors. All patients received kidneys from living donors and were given the same immunosuppression protocol (PRED plus CSA). Donor glomerular filtration rate (GFR) was determined before nephrectomy and at a mean period of 30 (range 10-50) months after nephrectomy. The graft GFR was measured at 1, 3, 6, and 12 months and at the most recent follow-up visit. Moreover, the functional reserve of the graft was assessed by infusion of dopamine and an amino acid. The postnephrectomy GFR of donors in groups A and B were 74 +/- 18 and 72 +/- 20 ml/min/1.73 m2, respectively. The GFR of pediatric recipients was significantly lower than that of adult recipients at corresponding time points along the course of follow-up. The mean values of graft GFR were 47.6 +/- 20 and 63.8 +/- 29.6 ml/min/1.73 m2 for pediatric and adult recipients, respectively (P < 0.001). Moreover, the graft functional reserve was significantly lower in pediatric recipients. These data demonstrate that adult kidneys transplanted into pediatric recipients have lower GFR than those transplanted into adults, despite corrections for body surface area. Although the reason for this phenomenon is unknown, the observation may have important implications for management of pediatric recipients.  相似文献   

3.
BACKGROUND: A total of 110 patients, in whom kidneys from 95 living related and 15 cadaver donor, had experienced renal transplantation between February 1985 and October 1996 in our clinic. This study was conducted to evaluate the influence of the various pre-operative factors to the graft survivals and clinical course of patients in living related renal transplantation. METHODS: In 95 recipients, 17 adult patients had long term graft survivals over 5 years including 6 recurrent or denovo nephritis without chronic allografts nephropathy. Eight failed to graft loss attributed to chronic allografts nephropathy diagnosed within 5 years. Retrospective analysis were performed to elucidate the differences of these recipients. RESULTS: Donors of long graft survival recipients were younger (49.1 +/- 12.1 v.s. 58.9 +/- 10. 2) and had a better renal function evaluated by preoperative creatinine clearance in living related donors (115.5 +/- 37.0 v.s. 79.7 +/- 22.0 1/day). Graft long survival recipients had experienced less frequencies of acute rejection within 6 months (0.53 +/- 0.62: 8 patients, 9 times) compared with chronic allografts nephropathy recipients (1.00 +/- 0.53: 7 patients, 8 times). Long graft survival recipients had better responses to the antirejection therapy. Additionally acute rejection over 6 months were experienced only in chronic allografts nephropathy recipients. Higher serum creatinine level was revealed in recipients with chronic allografts nephropathy at 1 year after transplantation (1.27 +/- 0.27 v.s. 1.88 +/- 0.42 mg/dl). CONCLUSIONS: We concluded that donor age and renal function are related to the graft long survival as background factors. Long graft survival recipients had less frequency of acute rejection and good response to the antirejection therapy. In recipients with of acute rejection and good response to the antirejection therapy. In recipients with chronic allografts nephropathy, serum cretine level had already increased gradually within 1 year.  相似文献   

4.
BACKGROUND: To increase the utilization of cadaveric donor kidneys, we have recently expanded our acceptable criteria to include aged donors (frequently with a history of hypertension), by selectively using both donor kidneys (dual transplant) into a single recipient. METHODS: To define when these expanded criteria donor (ECD) kidneys should be used as a single versus a dual kidney transplant, we retrospectively reviewed 52 recipients of ECD kidneys that had been turned down by all other local centers between 1/1/95 and 11/15/96. Fifteen patients received dual transplants, whereas the remaining 37 received single kidneys. Of the dual kidney recipients, 14 of 15 ECD were > or = 59 years of age, 10 of 15 were hypertensive, and 9 of 15 were both. Of the single recipients, 11 of 37 ECD were > or = 59 years of age, 11 of 37 were hypertensive, and 7 of 37 were both. All patients received cyclosporine-based triple-drug therapy. We compared seven donor (D) and sixteen recipient outcome variables in single versus dual kidney transplants as subgrouped by: (1) donor admission creatinine clearance (D-AdC(Cr)) < 90 ml/min; (2) D-age > or = 59 years; and (3) cold storage (Cld Stg) < or > 24 hr. RESULTS: In the group with D-AdC(Cr) < 90, there was a significantly higher incidence of delayed graft function (DGF) in single versus dual recipients (9 of 20 [45%] vs. 1 of 11 [9%]; P=0.04) and worse early graft function based upon mean serum creatinine at 1 and 4 weeks (5.3+/-3.3 and 2.8+/-2.0 vs. 1.7+/-0.6 and 1.4+/-0.5 mg] dl; P<0.05). In the group with D-age > or = 59, recipients of single kidneys had significantly higher mean serum creatinine at 1, 4, and 12 weeks versus recipients of dual kidneys (5.1+/-3.3, 3.4+/-2.1, 2.8+/-1.5 versus 2.8+/-2.5, 1.5+/-0.6, 1.6+/-0.5 mg/dl; P<0.05). Cld Stg time also had an impact on DGF and early outcome. Recipients of dual kidneys stored less than 24 hr had a significantly lower incidence of DGF versus single kidneys stored more than 24 hr (10% vs. 46%; P<0.05) and better early graft function based on mean serum creatinine at 1, 4, and 12 weeks (1.9+/-0.8, 1.3+/-0.4, 1.5+/-0.2 vs. 6.6+/-3.4, 3.0+/-1.6, 2.9+/-1.9 mg/dl; P<0.05). The overall 1-year patient and graft survivals were 96% and 81% vs. 93% and 87% (P=NS) in recipients of single ECD versus dual ECD kidneys. CONCLUSIONS: In conclusion, we believe that kidneys from ECD with D-AdC(Cr) < 90 ml/min and D-age > or = 59 should be used as dual kidney transplants, keeping the Cld Stg time at < 24 hr to minimize the effect of Cld Stg on early graft function.  相似文献   

5.
OBJECTIVE: Mature lobar transplantation will increase the pediatric donor organ pool, but it remains unknown whether such grafts will grow in a developing recipient and provide adequate long-term support. We hypothesized that a mature pulmonary lobar allograft implanted in an immature recipient would grow. METHODS: We investigated our hypothesis in a porcine orthotopic left lung transplant model using animals matched by the major histocompatibility complex to minimize the effects of chronic rejection. Twenty-three immature animals (< 12 weeks of age and < 10 kg total body weight) received either sham left thoracotomy (SH control, n = 4), left upper lobectomy to study compensatory growth (UL control, n = 4), age-matched immature whole left lung transplants (IWL TXP, n = 6), mature (donor > 1 yr in age and > 40 kg in total body weight) left lower lobe transplants (MLL TXP, n = 5), or mature left upper lobe transplants (MUL TXP, n = 4). Twelve weeks after implantation, functional residual capacity of the left lung was measured and arterial blood gas samples were obtained after the native right lung had been excluded. The graft was excised and weighed, and samples for microscopy and wet/dry ratios were collected. RESULTS: Initial and final graft weights were as follows: IWL TXP group (34.6 +/- 1.5 and 107.8 +/- 5.9 gm, p < 0.0001), MLL TXP group (72.4 +/- 6.8 and 111.4 +/- 8.7, p < 0.001), and MUL TXP group (32.8 +/- 1.3 and 92.8 +/- 7.1 gm, respectively, p < 0.004). No significant differences between groups were demonstrated when functional residual capacity, wet/dry ratios, or oxygenation were compared. Immunohistochemical staining for the nuclear antigen Ki-67 demonstrated dividing pneumocytes. CONCLUSIONS: We conclude that a mature lobar graft implanted into an immature recipient grows by pneumocyte division in this model. Mature lobar transplants can be expected to grow and provide adequate long-term function in developing recipients.  相似文献   

6.
BACKGROUND: To maximize the renal donor pool, cadaveric pediatric en bloc kidneys have been transplanted as a dual unit by some transplant centers. We compared the short- and long-term outcomes of adult recipients of cadaveric pediatric en bloc renal transplants versus those of matched recipients of cadaveric adult kidneys. METHODS: Thirty-three adults who received pediatric en bloc kidney transplants between April 1990 and September 1997 were retrospectively identified and were compared with 33 matched adults who received adult cadaveric kidney transplants. The groups were identical for transplantation era, immunosuppression, recipient sex, race, cause of renal failure, mean weight, and follow-up duration (37.8 vs. 37.5 months). The mean recipient age study versus control was lower (36.3 vs. 48.9 years, P=0.0003). Results. There was no difference between the en bloc and adult donor groups in the 3-year patient survival rates (95% vs. 87%, P=0.16) or the 3-year graft survival rates (87.3% vs. 84.2%, P=0.35). Further, there was no difference in en bloc patient or en bloc graft survival time stratified by recipient age (14-44 vs. >45 years, P=0.11), en bloc donor age (<24 vs. >24 months, P=0.39), or recipient weight (<60, 61-75, >75 kg; P=0.60). Differences in serum creatinine (mg/dl) for the en bloc versus the control group at the time of discharge (3.0 vs. 7.8 mg/dl, P=0.06), at 1 year (1.4 vs. 2.0 mg/dl, P=0.06), and at 2 years (1.1 vs. 1.6 mg/dl, P=0.14) had dissipated by the time of the 5-year follow-up examination (1.1 vs. 1.6 mg/dl, P=0.14). Vascular complications were more prevalent in the en bloc group: renal vein thrombosis (one case), thrombosis of donor aorta (two cases), arterial thrombosis of one renal moiety (two cases), and renal artery stenosis (two cases). There were no differences between groups in delayed graft function, acute or chronic rejection, posttransplant hypertension, posttransplant protein-uria, or long-term graft function. CONCLUSIONS: Collectively, these data indicate that transplanting pediatric en bloc kidneys into adult recipients results in equivalent patient and graft survival compared with adult cadaveric kidneys. Further, the data also suggest that pediatric en bloc kidneys need not be strictly allocated based on recipient weight or age criteria.  相似文献   

7.
BACKGROUND: Ischemia-reperfusion injury after organ transplantation is a major cause of delayed graft function. We showed earlier that antisense oligodesoxynucleotides (ODN) for intercellular adhesion molecule-1 (ICAM-1) ameliorate reperfusion injury after acute ischemia. This study tested the hypothesis that antisense ODN for ICAM-1 prevents ischemia-reperfusion injury and facilitates immediate graft function in a rat autotransplantation model. METHODS: Both kidneys were removed from male Lewis rats and re-implanted the left kidney after 30 minutes of cold ischemia time. The warm ischemia time was 60 minutes. Sham operated, uninephrectomized animals served as controls for renal function and histology. ICAM-1 antisense ODN (5 mg/kg), reverse ODN, or saline-vehicle were administered to donor animals i.v. six hours before autotransplantation. Glomerular filtration rate (insulin clearance), and serum creatinine concentrations were measured 24 hours post-transplantation. Tubular necrosis severity was assessed by histological grading scale. ICAM-1 expression was determined by immunohistochemistry and Western blot. RESULTS: Antisense ODN decreased ICAM-1 expression and leukocyte infiltration significant. Antisense ODN-treated animals showed significantly less tubular necrosis, than controls. Serum creatinine of antisense ODN-treated animals (N = 6) was 0.55 +/- 0.02 mg/dl compared to 1.92 +/- 0.07 mg/dl in reverse ODN-treated controls (N = 6; P < 0.01), 24 hours after transplantation. Antisense ODN-treated animals had normal GFR (0.93 +/- 0.07 ml/min/kidney wt) compared to sham-operated animals (0.95 +/- 0.09 ml/min/kidney wt), while autotransplanted animals treated with reverse ODN or saline-vehicle were all anuric. The ischemia-reperfusion-induced up-regulation of MHC class II was totally prevented by antisense ODN. CONCLUSIONS: ICAM-1 inhibition ameliorates ischemia-reperfusion injury and prevents delayed graft function. Antisense ODN-treatment of donors or donor organs for ICAM-1 may be useful for the prevention of reperfusion injury in human renal transplantation.  相似文献   

8.
BACKGROUND: Traditionally, elderly donor kidneys have not been widely accepted for transplantation on the assumption of inferior performance. However, the United Network for Organ Sharing reports an increase in the number of elderly donors from less than 2% in 1982 to 24% in 1995. This trend is commensurate with the increase of older dialysis patients and an overall increase in the elderly population in the United States (1). Optimal utilization of these kidneys is essential to overcome the acute organ shortage. METHODS: In this study, we transplanted 25 kidneys from elderly donors (ages 56-72 years) into young adult recipients (ages 20-50 years) (group 1) over a 4-year period. We compared the results with matched recipients of young adult donor kidneys (group 2) with regard to long-term kidney function and graft survival. A pretransplant biopsy of elderly donor kidneys was carried out and a frozen section report was obtained. Only those kidneys showing glomerulosclerosis of less than 20% were accepted for transplantation. All cadaveric kidneys were preserved in University of Wisconsin solution. RESULTS: Pretransplant biopsies of elderly donor kidneys showed structural deficits, which included glomerulosclerosis in 85%, arteriolar and/or mesangial thickening in 75%, and interstitial lymphocyte infiltration in 30%. The mean serum creatinine was 2.4+/-0.74, 2.2+/-0.56, and 2.9+/-0.76 mg/100 ml in group 1 and 1.5+/-0.55, 2.3+/-2.24, and 1.7+/-0.62 in group 2 at 1, 3, and 5 years, respectively. The patient survival was 92%, 92%, and 88% in group 1, and 100%, 100%, and 100% in group 2 at 1, 3, and 5 years, respectively. The graft survival was 80%, 64%, and 56% in group 1 and 100%, 96%, and 88% in group 2 at similar time intervals. The differences in the serum creatinine and graft survival between the two groups were statistically significant (P < 0.05). CONCLUSIONS: Most of the elderly donor kidneys with structural deficits transplanted into young adults provided suboptimal function and inferior long-term graft survival. To maximize the utilization and optimize the survival of elderly donor kidneys, we propose transplantation of these kidneys into age-matched recipients with similar physiological requirements as those of donors, with regard to kidney function.  相似文献   

9.
BACKGROUND: Recipient hepatitis C virus (HCV) seropositivity has been associated with inferior outcomes in renal transplantation (RTx). We sought to determine whether donor HCV+ status influenced the incidence of rejection, liver dysfunction, and graft survival in HCV+ recipients. METHODS: We reviewed 44 HCV+ recipients (R+) receiving RTx from HCV+ (D+) and HCV- (D-) donors between February 1991 and September 1996. All patients were followed to the end of the study period (mean=36 months, range=12-60 months). We compared the R+ group with a demographically matched cohort of 44 HCV- recipients (R-). RESULTS: Of the 44 R+, 25 (57%) had a total of 48 rejection episodes. Among the 44 R-, 32 (73%) had 58 rejection episodes (P>0.1). Within the R+ group, 28 were D+/R+; of these 14 (50%) had 27 rejection episodes, whereas among the 16 D-/R+, 11 (68%) had 21 rejection episodes (P>0.3). Graft and patient survival was similar in both the groups (86.4% and 91%, respectively). Liver dysfunction was slightly increased in the R+ group (4/44 vs. 0/44, P>0.1), with one death due to liver failure in this group. CONCLUSION: Donor HCV+ status had no influence on outcomes in HCV+ recipients after kidney transplantation in the short term. The incidence of rejection, graft loss, and mortality was comparable between the D+/R+ and D-/R+ groups. Furthermore, rejection, graft loss, and death were identical in R+ and R-groups throughout the 5-year study period. We therefore conclude that HCV+ recipients can safely receive kidney transplants without concern about donor HCV status or fear of adverse events from their own HCV+ status.  相似文献   

10.
BACKGROUND: Renal transplantation in infants is frequently complicated by graft thrombosis and accelerated rejection reactions. We herein tested the hypothesis that the amount of blood required to sustain normal perfusion of an adult renal allograft transplanted into a pediatric recipient would surpass the cardiac output and aortic blood flow of the recipient and that the ensuing low flow in full-size grafts (FSG) would induce a release of thrombogenic substances. METHODS: In a porcine renal transplant model, adult FSG were transplanted into pediatric recipients. Macro- and microhemodynamic as well as metabolic data were recorded. Surgically size-reduced grafts (RSG) served as controls. RESULTS: Donor weight was 55.1+/-4.8 kg and 9.6+/-0.9 kg for recipients. FSG weight was 122+/-16 g and 65+/-14 g for RSG. Blood flow in donor kidneys was 20% higher than the infrarenal aortic blood flow of recipients. After reperfusion, mean arterial pressure in recipients of FSG but not RSG dropped to 64 mmHg, despite an increase in cardiac output by 60%. FSG but not RSG were polyuric and proteinuric. The release of endothelin and thromboxane B2 into the circulation was higher from FSG when compared with RSG (P<0.05 for endothelin after 60 min; NS for thromboxane B2). CONCLUSIONS: After transplantation of FSG into pediatric recipients, the macrohemodynamic limitations of the recipient cause microcirculatory disturbances in the graft, which contribute to the release of vasoconstrictive and prothrombotic substances and an impaired early graft function. Some of those effects can be ameliorated by surgically size reducing the renal graft.  相似文献   

11.
BACKGROUND: We previously reported excellent outcome at 6 months after transplantation in recipients of expanded criteria donor kidneys that other local centers had declined, kidneys that nobody wanted (KNW), versus controls. We now report follow-up after 23 months. METHODS: We retrospectively reviewed 27 donor and 24 recipient characteristics in 126 adult recipients of transplants from January 1, 1995, to November 25, 1996. RESULTS: Donors of control kidneys versus KNW were younger and had significantly higher minimum 4-hr urine output. Recipients of control kidneys versus KNW had significantly more HLA matches and lower 3-month posttransplant serum creatinine levels. Patient and graft survival rates were similar between the control kidneys versus the KNW. We also compared the control kidneys and KNW with regard to prompt function or delayed graft function and satisfactory versus unsatisfactory function (unsatisfactory: serum creatinine > or =2.5 ml/dl or graft loss at 6 months) to identify donor and recipient characteristics associated with delayed graft function and unsatisfactory outcome. The incidence of rejection was significantly lower in control kidneys and KNW with satisfactory function versus control kidneys and KNW with unsatisfactory function. CONCLUSIONS: These data demonstrate: (1) similar graft survival at 12 months, (2) lower donor age, (3) higher minimum 4-hr urine output, and (4) more HLA matches in recipients of control kidneys versus KNW. Optimal outcome was achieved in recipients of control kidneys and KNW with prompt function and satisfactory function based upon serum creatinine in the first 6 months and in recipients with lower rates of rejection. Although outcome is dependent upon many donor and recipient variables, we believe that with careful donor and recipient selection, excellent outcome can be achieved using expanded criteria donor kidneys.  相似文献   

12.
BACKGROUND: In 1994, a policy of double renal allografting (DUAL) was used at two centers within our local organ procurement organization to increase utilization of kidneys from older donors that would otherwise be discarded. Both kidneys from an older donor (age > 60 years) were selectively transplanted into a single adult recipient. METHODS: The relative impact of a DUAL policy on the utilization of older donor kidneys is examined for the period of April 1994 to April 1996. Actual utilization is compared with the hypothetical case in which a DUAL policy is not present. RESULTS: In the actual setting, a total of 75 kidneys from older donors (> 60 years) were accepted for transplantation. Thirty-six kidneys were transplanted as singlets, and 16 additional kidneys were transplanted as DUAL renal allografts. Thus, a 44% increase in transplantable kidneys, and a 22% increase in patients transplanted with kidneys from older donors, was realized. In the actual setting, 23 older kidneys were discarded; without the DUAL policy, 39 kidneys would have been deemed untransplantable. When compared with the actual (n = 52) and hypothetical number of kidneys transplanted without a policy of DUAL transplantation (n = 36), the DUAL policy significantly increased the utilization of older donor kidneys (P = 0.01). The actuarial 1-year graft survival rate of the dual kidneys was 100%, with a mean follow-up of 11.1 +/- 2.9 months. Mean 6-month and 1-year serum creatinine level were 1.76 +/- 0.4 and 1.63 +/- 0.6 mg/dl, respectively. CONCLUSIONS: A DUAL policy significantly increased the number of older donor kidneys transplanted in a single organ procurement organization and reduced the rate of discard of older donor kidneys over a 2-year period. Long-term follow-up is necessary to substantiate satisfactory graft function in DUAL from older donors.  相似文献   

13.
AIMS: The aims of this study were to examine the influence of advanced recipient and donor age on the long term outcome of corneal transplantation. METHODS: Records of 1036 penetrating corneal grafts in recipients aged > or = 80 years at surgery (defined as the elderly subset) and 8092 donor corneas used for transplantation were obtained from the Australian Corneal Graft Register database, Kaplan-Meier graft survival plots were compared using log rank statistics. RESULTS: Elderly recipients constituted 15% of the recipient pool. The major indication for corneal transplantation in the elderly was bullous keratopathy. Graft survival fell with increasing recipient age (p < 0.00001); the major cause of graft failure was rejection (33%). The desired outcome in 51% of cases was to improve vision and in 42% of cases to relieve pain; 23% of elderly recipients achieved a Snellen acuity of 6/18 or better in the grafted eye and 66% recorded improved acuity after transplantation. Elderly recipients suffered more complications and comorbidities in the grafted eye than did younger recipients. Donor age (stratified in 10 year intervals) did not influence corneal graft survival significantly (p = 0.10). CONCLUSIONS: Elderly graft recipients fared less well after corneal transplantation than did younger recipients, but outcomes in terms of long term graft survival and visual rehabilitation were still good. Donor age did not affect graft survival.  相似文献   

14.
BACKGROUND: Due to their vasodilatory effect, calcium antagonist may have a renoprotective against cyclosporin (CsA)-induced nephrotoxicity and rise in blood pressure (BP) seen in renal transplantation. METHODS: In order to evaluate the effect of the calcium antagonist felodipine on renal function and BP during cyclosporin treatment, 79 CsA-treated renal transplant recipients were investigated during the first 3 months after transplantation in a randomized, double-blind, placebo-controlled study with two parallel groups. Felodipine (ER tablets, 10 mg) or placebo was given prior to transplantation and each day during the study period. The patients were assessed twice, i.e. at 4-6 weeks and at 10-12 weeks after transplantation. Renal plasma flow (RPF) and glomerular filtration rate (GFR) were measured by constant infusion technique. Tubular function was estimated from clearance of lithium. RESULTS: At 6 weeks after transplantation, felodipine caused a significantly higher RPF [felodipine: 219 +/- 70 ml/min; placebo: 182+/-56 ml/min (mean+/-1 SD); P=0.03]. No differences were found in GFR, filtration fraction (FF), tubular sodium handling, or sodium excretion. Felodipine lowered BP significantly. At 12 weeks after transplantation, felodipine caused a significantly higher GFR (felodipine: 49+/-18 ml/min; placebo: 40+/-16 ml/min; P=0.05) and RPF (felodipine: 225+/-77 ml/min; placebo: 175+/-48 ml/min; P<0.01). No difference was found in FF. Felodipine lowered BP significantly. No differences were found with regard to duration of primary anuria, hospitalization time, number of rejection episodes, plasma creatinine day 7 post-transplant, or treatment doses of CsA. CONCLUSIONS: It is concluded that in renal transplant recipients treated with CsA, felodipine significantly increased both GFR and RPF 3 months after transplantation when compared with placebo, despite a concomitant lowering of BP. A possible antagonizing affect of felodipine against CsA-induced nephrotoxicity in these patients is suggested.  相似文献   

15.
To study the role of IL-4 in tolerance induction and transplant arteriosclerosis, BALB/c hearts were transplanted into C57BL/6J wild-type or IL-4 knockout (IL-4(-/-)) recipients. A 30-day course of anti-CD4/8 mAb was used to induce long term graft survival. Primary graft survival was 50% (5 of 10) in IL-4(-/-) recipients comparable to 63% (5 of 8) in wild-type recipients. Mice with allografts surviving >80 days were tested for tolerance by challenge with a second donor or third party (CBA) heart. Secondary donor-strain heart grafts survived >30 days, but showed histologic evidence of ongoing alloimmune response. Third party hearts rejected rapidly. Although immunostaining and 32P RT-PCR assays showed no differences in the mononuclear cell infiltration and T cell activation between IL-4(-/-) and wild-type tolerant recipients, some monokines (IL-12, TNF-alpha, and allograft inflammatory factor-1) were up-regulated in grafts from IL-4(-/-) recipients. Computer-assisted analysis of elastin-stained vessels revealed that the severity of vascular thickening (percentage of luminal occlusion, mean +/- SD, n = 329) was similar in grafts from IL-4(-/-) (63.7 +/- 16.9%) and wild-type (69.5 +/- 17.6%) recipients. Thus, IL-4 deficiency did not alter primary or secondary graft survival, infiltration, or vascular thickening. The selective alterations in monokine expression suggests that alternative pathways are activated and may compensate in IL-4(-/-) mice.  相似文献   

16.
BACKGROUND: Failure of the donor (graft) heart is the main cause of mortality in the first month after orthotopic cardiac transplantation. In a preliminary study marked downregulation of cardiac beta adrenoceptor density was found in apparently normal donor hearts of recipients who developed severe cardiac failure soon after implantation. Cardiac beta adrenoceptors are an important factor in the development of cardiac failure in the human heart. The aim of this study therefore was to determine whether fatal graft failure in the first month after transplantation is associated with downregulation of beta adrenoceptor density in the donor heart. PATIENTS AND METHODS: Right ventricular endomyocardial biopsy specimens were taken from consecutive adult donor patients immediately before implantation. A previously described radioligand binding method was used to determine beta adrenoceptor density in consecutive patients who developed fatal graft failure and died within 1 month of transplantation and in a group of control donors transplanted during the same period. RESULTS: Perioperative fatal graft failure developed in 13 patients. Forty one specimens from donor hearts that were transplanted into recipients who did not develop fatal graft heart failure formed the control group. There were no systematic differences in donor or recipient characteristics between the graft heart failure and control groups. In particular donor catecholamine requirement and recipient pulmonary vascular resistance did not differ between groups. Total beta adrenoceptor density was reduced in the fatal graft heart failure group compared with that in the controls (13.4 (7) fmol/mg v 21 (7) fmol/mg; P < 0.01). There was a positive correlation between beta adrenoceptor density in the donor heart and time to death in the graft heart failure group (r2 = 0.3, P < 0.05). The beta adrenoceptor binding affinity (Kd) did not differ between the graft failure group and the controls (47 (6) pM v 44 (7) pM). CONCLUSION: The development of perioperative fatal cardiac failure after orthotopic cardiac transplantation is associated with downregulation of beta adrenoceptors in the donor heart before implantation.  相似文献   

17.
AIM: Renal functional reserve (RFR), resulting from an increase in glomerular filtration (GFR) after protein load, is a matter of debate. In kidney transplant recipients most studies have failed to show conclusive results, reporting either the absence, the reduction or the presence of renal reserve in normo-functioning kidneys. The aim of this study was to investigate RFR in kidney transplant patients as well as the possible hormonal vasoactive alterations underlying the reduction of renal reserve reported in some patients. PATIENTS AND METHODS: We studied 8 controls and 25 patients, the latter with no history of acute rejection for at least 12 months and GFR >50 ml/min. The 25 patients were divided into 2 groups based on the presence (10) or the absence (15) of RFR. RESULTS: Both the RFR group and the controls experienced a similar increase of GFR after oral protein load: 24.3 +/- 15.57% vs 24.4 +/- 10.8%. The group without RFR showed a paradoxical reduction of GFR after oral protein load: 13.3 +/- 13.2% (p <0.001). We analyzed the filtration fraction (FF) and observed that the group without RFR had higher values than the group with RFR and the controls: 0.35 +/- 0.11 vs 0.29 +/- 0.07 (p = 0.01) and vs 0.26 +/- 0.02 (p = 0.04). The hyperfiltration state observed in the group without RFR was sustained by a high level of thromboxane. The urine ratio TxB2/6ketoPgF1alpha was higher in the group without RFR than in the RFR group 0.78 +/- 0.2 vs 0.64 +/- 0.1 (p = 0.01). This ratio decreased only in the RFR group after a meat meal. In all the patients, changes of TxB2/6ketoPGF1alpha were inversely correlated to changes of GFR after a meat meal (r = -0.6, p = 0.01). CONCLUSIONS: In conclusion, these data demonstrate that kidney transplant recipients with good organ function can be grouped according to the presence of RFR. RFR appears to be inversely correlated with the TxB2/6ketoPGF1alpha ratio, and its decrease seems to be linked to the failure of thromboxane to decrease and prostacycline to increase after a meat meal.  相似文献   

18.
BACKGROUND: Outcome after renal transplantation in children has been variable. We undertook a retrospective study of our experience over the past five years. STUDY DESIGN: From January 1, 1988, to October 15, 1992, 60 renal transplantations were performed upon 59 children at the Children's Hospital of Pittsburgh. Twenty-eight (47 percent) of the kidneys were from cadaveric donors, and 32 (53 percent) were from living donors. The recipients ranged in age from 0.8 to 17.4 years, with a mean of 9.8 +/- 4.8 years. Forty-six (77 percent) recipients were undergoing a first transplant, while 14 (23 percent) received a second or third transplant. Eight (13 percent) of the patients were sensitized, with a panel reactive antibody of more than 40 percent. Eleven of the 14 patients undergoing retransplantation and seven of the eight patients who were sensitized received kidneys from cadaveric donors. Thirty-three (55 percent) patients received cyclosporine-based immunosuppression, and 27 (45 percent) received FK506 as the primary immunosuppressive agent. RESULTS: The median follow-up period was 36 months, with a range of six to 63 months. The one- and four-year actuarial patient survival rate was 100 and 98 percent. The one- and four-year actuarial graft survival rate was 98 and 83 percent. For living donor recipients, the one- and four-year actuarial patient survival rate was 100 and 100 percent; for cadaveric recipients, it was 100 and 96 percent. Corresponding one- and four-year actuarial graft survival rates were 100 and 95 percent for the living donor recipients and 96 and 69 percent for the cadaveric recipients. Patients on cyclosporine had a one- and four-year patient survival rate of 100 and 97 percent, and patients on FK506 had a one- and three-year patient survival rate of 100 and 100 percent. Corresponding one- and four-year actuarial graft survival rates were 100 and 85 percent in the cyclosporine group, while one- and three-year actuarial graft survival rates were 96 and 84 percent in the FK506 group. The mean serum creatinine level was 1.24 +/- 0.64 mg per dL; the blood urea nitrogen level was 26 +/- 13 mg per dL. The incidence of rejection was 47 percent; 75 percent of the rejections were steroid-responsive. The incidence of cytomegalovirus was 10 percent. The incidence of post-transplant lymphoproliferative disorder was 8 percent. None of the patients on cyclosporine were able to be taken off prednisone; 56 percent of the patients receiving FK506 were taken off prednisone successfully. Early growth and development data suggest that the patients receiving FK506 off prednisone had significant gains in growth. CONCLUSIONS: These results support the idea that renal transplantation is a successful therapy for end-stage renal disease in children. They also illustrate the potential benefits of a new immunosuppressive agent, FK506.  相似文献   

19.
BACKGROUND: Graft coronary artery disease (CAD) is an increasingly important problem during long-term survival after heart transplantation, but the importance of cellular rejection, in particular late after transplantation, remains undetermined. METHODS and RESULTS: We analyzed 492 coronary angiographies (967+/-705 days after transplantation; range, 49 days to 9.4 years) and 5201 endomyocardial biopsies (518+/-648 days after transplantation) from 156 patients (age, 47+/-11 years). Patients with angiographically detectable graft CAD had significantly more episodes of rejection requiring augmentation of immunosuppressive therapy (i.e., International Society of Heart and Lung Transplantation score > or = 3A) than those without graft CAD during the first (3.7+/-2.6 vs. 2.2+/-2.0, P<0.001) as well as subsequent years after transplantation (1.2+/-1.9 vs. 0.4+/-0.9, P<0.01). Multivariate logistic regression analysis including established risk factors for CAD, ischemic time, gender and age of donors and recipients, number of mismatches, cytomegalovirus infection, and drug therapy showed that the number of rejections during the first [odds ratio (OR)=1.39, P<0.005] as well as subsequent years (OR=1.49, P<0.05), previous cytomegalovirus infection (OR=3.21, P<0.05), donor age >40 years (OR=2.97, P<0.05), and current or former smoker status (OR=2.76, P<0.05) were independent predictors of graft CAD. In patients without angiographically detectable graft CAD 1 year after transplantation, the number of rejections after the first year was even more strongly related to graft coronary artery disease than in the total patient population, underlining the importance of late cellular rejection (OR=1.74, P<0.005). CONCLUSION: Rejection requiring augmentation of immunosuppression early and late after transplantation is an independent risk factor for the development of angiographically detectable graft CAD. Hence, the search for and treatment of moderate or severe rejection seems to be prudent even late after transplantation.  相似文献   

20.
Neonatal skin allografts can be tolerogenic when transplanted to appropriately immunosuppressed hosts. Single grafts of neonatal skin survive longer than adult skin grafts when recipients are treated with antilymphocyte serum (ALS) and donor bone marrow cells (BMC). Neonatal skin grafts can also prolong the survival of adult grafts of the same donor strain simultaneously cotransplanted with the neonatal grafts. To probe the mechanisms involved in this cotransplantation phenomenon, we delayed placement of the neonatal cotransplants relative to grafting with adult skin. Neonatal allografts placed either 7-9 days or 14 days after grafting with adult skin significantly prolonged adult graft survival in mice treated with ALS and BMC. However, day 0-placed neonatal cotransplants must remain on the recipient for > 2 weeks to prolong adult graft survival. Removal of cotransplants from ALS- and BMC-treated recipients after 7 or 14 days abrogated the cotransplantation effect. If left in place until day 21, neonatal cotransplants could significantly prolong adult graft survival, but did not induce the long-term graft survival observed in approximately 50% of the recipients whose cotransplants were not removed. Cotransplant removal after 1 year did not affect subsequent adult graft survival. Additionally, cotransplants were removed from recipients either on day 14 or from long-term graft-bearing mice and retransplanted to other ALS/BMC-treated recipients. These retransplanted grafts were unable to prolong survival of adult grafts on the new recipients. After transplant, but not before transplant, cyclophosphamide treatment of recipients prevented expression of the cotransplant effect in ALS-treated mice. However, recipient splenectomy > or = 1 week before grafting did not interfere with the effect. These results reflect on the contributions of the donor tissue, and the recipients' response, to the tolerogenic signals that permit a neonatal cotransplant to prolong adult graft survival.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号