FRONTIERS IN ENDOSCOPY, SERIES # 15

Bile Duct Injuries: Multidisciplinary Evaluation and Treatment

Read Article

Here we discuss bile duct injuries, which can have substantial medical and psychological impact on patients and can require significant healthcare resources for diagnosis and management. They are most commonly treated using endoscopic techniques, but may require a multi-disciplinary approach with percutaneous and surgical approaches as well. The topic of bile duct injuries pertains to several medical specialties including general and transplant surgery, gastroenterology, interventional radiology, and general medicine services which may work as a multidisciplinary team to formulate an individualized plan for each patient.

Tyler R. McVay MD, University of Utah School of Medicine, Gastroenterology and Hepatology. Douglas G. Adler MD, FACG, AGAF, FASGE, Associate Professor of Medicine, Director of Therapeutic Endoscopy, Director, GI Fellowship Program, Gastroenterology and Hepatology, University of Utah School of Medicine, Huntsman Cancer Center, Salt Lake City, UT

INTRODUCTION

Bile duct injuries can have substantial medical and psychological impact on patients and can require significant healthcare resources for diagnosis and management. Iatrogenic injury during surgery is the most common etiology, but both blunt and penetrating trauma can cause bile duct injuries as well. There is a significant prevalence of variant biliary anatomy that poses a risk for biliary injury during surgery. Bile duct injuries are most commonly treating using endoscopic techniques but may require a multi- disciplinary approach with percutaneous and surgical approaches as well. The topic of bile duct injuries pertains to several medical specialties including general and transplant surgery, gastroenterology, interventional radiology, and general medicine services which may work as a multidisciplinary team to formulate an individualized plan for each patient.

CAUSES OF BILE DUCT INJURIES

Laparoscopic Cholecystectomy

Bile duct injuries most commonly occur during surgery, and the surgery most frequently implicated is laparoscopic cholecystectomy. Bile ducts can be accidentally clipped, lacerated, avulsed, or completely transected during surgery, with the most severe injuries being the least common. Cystic duct remnant injuries have been and remain the most common injury following cholecystectomy. (Figure 1) The laparoscopic approach to cholecystectomy was introduced in the late 1980s, and although it provides many advantages over open cholecystectomy, it has been associated with a higher rate of biliary complications.1,2,3,4,5,6 The rate of bile duct injury for open cholecystectomy has been reported to be 0.1-0.2% (although it has been reported to be as high as 0.75%).2,3 Many large population-based studies have estimated that the rate of bile duct injury for laparoscopic cholecystectomy is 0.3-0.5%.1,2,4,5,6

Due to intra-operative difficulties, surgeons convert about 4.7-8.2% of laparoscopic cholecystectomies to open procedures. The principal reason for converting in these cases is poor or limited visualization, distorted anatomy due to inflammation, and to control bleeding.4 Patients with a history of upper abdominal surgery often have distorted anatomy as well and are more likely to require conversion to open cholecystectomy.5 There is a learning curve associated with the technical skill required to perform laparoscopic cholecystectomies. One study found that novice surgeons who have performed less than ten laparoscopic cholecystectomies endure a bile duct injury rate of 0.49%, but surgeons with a tally greater than a hundred have a rate of only 0.04%.5

A newer laparoscopic approach, known as single- port laparoscopic cholecystectomy, is typically performed through a single incision site in the umbilicus and poses additional opportunities and challenges for surgeons. This new approach is less invasive and potentially results in quicker recovery time, less post- operative pain, and better cosmetic results for patients.7,8 A recent meta-analysis of eleven randomized controlled trials revealed a nonsignificant difference in the bile duct injury rate comparing the single-port technique (0.4%) with conventional laparoscopic cholecystectomy (0%) with use of three or four ports.8 However, the authors recognized that only two of the included trials in the meta-analysis offered evaluation of bile duct injuries, which limits the validity of the results.8

ORTHOTOPIC LIVER TRANSPLANT

Anastomotic Injuries

Liver transplant patients, both recipients and living donors, are at risk for bile duct injuries. Despite advances in surgical technique, biliary complications are still reported in 10-39% of liver transplant recipients, including strictures, leaks, anastomotic dehiscence, or a combination thereof. 9,10,11,12 Strictures can occur at the site of the bile duct anastomosis or at distant locations in the biliary tree, and bile leaks can result from a failed anastomosis or direct injury to bile ducts. The anastomosis of the bile ducts between the donor liver and recipient can occur via a duct-to- duct, end-to-end choledochocholedochostomy, or via hepaticojejunostomy, and strictures occur with any of these approaches.9,11,12 Anastomotic strictures can result from surgical technique or bile duct ischemia that results in fibrosis and stenosis of the anastomosis.9 (Figure 2) The risk of developing an anastomotic stricture increases with time from transplant surgery, with a rate of 6.6% at 1 year and 12.3% at 10 years.9

The following risk factors have been associated with the development of anastomotic strictures after orthotopic liver transplant: the presence of a post- operative bile leak, female donor-male recipient transplant combination, and a more recently performed transplant (although this may reflect improved diagnosis and increased graft survival which allows more time for anastomotic strictures to develop).9,11 A direct duct- to-duct anastomosis has been associated with a higher risk of post-transplant biliary complication than Roux- en-Y hepaticojejunostomy.12 Anastomotic strictures have a high recurrence rate of nearly 20% after initial successful treatment. Patients at risk for recurrent anastomotic strictures are those who experience a bile leak or have a longer time to presentation.13

Non-anastomotic Injuries

Non-anastomotic biliary strictures following orthotopic liver transplant can occur for several reasons. Bile duct ischemia from hepatic artery thrombosis is a recognized cause of non-anastomotic strictures.9 Ischemic-type biliary lesions, characterized by diffuse intrahepatic biliary strictures and dilatations in the absence of hepatic artery thrombosis, occur in 5-15% of orthotopic liver transplant patients and are not completely understood.14 Proposed risks for ischemic-type biliary lesions include compromised blood flow to the peri-biliary capillaries, immunologic injuries, and cytotoxic injuries from bile salts.14,15 Efforts to provide better perfusion of the small peri-biliary capillaries at the time of graft retrieval have resulted in a decreased incidence of ischemic-type biliary lesions.15 Other known causes of non-anastomotic strictures include ABO blood type incompatibility and immunologic rejection.9

Adult Living Donor Liver Transplant (ALDLT)

Approximately 1/3 of adult living donor liver transplantation recipients will develop some sort of biliary injury, although this rate has been reported in some centers to be as high as 67%.10,11 The incidences of bile duct leaks and strictures are similar in ALDLT recipients. For right hepatic lobe recipients, the anastomosis of the donor right hepatic duct to recipient common hepatic duct is associated with the lowest incidence of biliary complication when compared to all other anastomosis types.11 When donor liver grafts have more than one bile duct, a Roux-en-Y hepaticojejunostomy anastomosis is often used. Three- duct donor grafts have been associated with increased biliary complications when compared to single-duct donor grafts. The difference likely reflects a higher risk of bile duct ischemia when several small ducts are involved rather than a single large duct.11

The donors of ALDLT can also experience bile duct injuries. Bile leaks occur in 4-6% of living liver donors and are more common than bile strictures.16,17 Right lobe donors endure a higher bile leak rate and overall complication rate than left-sided donors.16,18,19 Right lobe grafts provide more liver parenchyma that can accommodate increased portal vein flow to the recipient, however this comes at the cost of a larger resection margin with increased risk of bile leak in the donor.16,20 A prospective study showed bile leaks in 7% of right hepatectomies and in 4% in left lateral hepatectomies and left lobe hepatectomies. Repeat operations are required in 1-4.5% of all living donors to address the most severe complications, but most bile leaks are controlled with less invasive techniques such as endoscopic retrograde cholangiography with stent placement, percutaneous transhepatic drainage or ultrasound-guided drainage.16,17,20

Transcatheter Arterial Chemoembolization (TACE) and Chemoinfusion (TACI)

Transcatheter arterial chemoembolization (TACE) and chemoinfusion (TACI) are interventional radiology procedures used to treat solid hepatic malignancies. These procedures are most often used to treat hepatocellular carcinoma (HCC) but are also used for other hepatic tumors as well. TACE consists of infusing iodized oil mixed with a chemotherapeutic agent directly into the tumor, followed by the infusion of gelatin sponge particles to embolize the blood vessels surrounding the tumor.21 Subcapsular bilomas and focal bile duct strictures are potential complications of TACE, and they often present in the context of serious bacterial infections such as cholangitis.21 Leaks and, more commonly, strictures in these patients can present in an acute or delayed manner.

Bile duct injury from TACE can occur in hepatic lobes separate from the tumor location. The iodized oil with the chemotherapeutic agent is often injected via the common hepatic artery, which can lead to bile duct necrosis with biloma or stricture formation in any segment of the liver.21 The hypertrophied peri-biliary capillaries in a cirrhotic liver may act as a vascular shunt that protects against bile duct ischemia after TACE. In fact, Child-Pugh class A patients have a higher rate of bile duct injury (15.2%) following TACE when compared with Child-Pugh class B and C patients (2.7%).22 Bile duct injury from TACE is also more common with non-hepatocellular carcinoma tumors (38.9%) compared to HCC tumors (11.3%).22 TACI, which involves chemoinfusion without embolization, has a significantly lower risk of bile duct injury than TACE.21

Resection and Radiation of Hepatic Tumors

Patients may undergo partial hepatectomy as definitive treatment for hepatocellular carcinoma. Bile leaks are common post-operative complications and occur in up to 12.8% of cases.23 Pre-operative treatments such as TACE or radiofrequency ablation do not increase the risk for development of a post-hepatectomy bile leak per se. However, if these pre-operative treatments were complicated by a bile leak or stricture, then the patient is at increased risk for a post-hepatectomy bile leak as well. Prolonged operative time has also been identified as a risk factor for bile leak after hepatectomy.23

Patients with either primary or metastatic liver tumors may receive radiation therapy as an element of their cancer treatment. Delivery of radiation to these tumors creates a risk of developing fibrosis and bile duct strictures. Stereotactic body radiation therapy (SBRT), which involves accurate delivery of radiation directly to the tumor, is a potential therapy for tumors in high-risk areas such as the hepatic hilum that cannot otherwise be treated safely with surgery, TACE, or radiofrequency ablation. When specific radiation doses are used (40 Gy or less), SBRT has been associated with minimal biliary toxicity and is a viable option for treatment of tumors within the hepatic hilum.24

Traumatic Bile Ducts Injuries

Penetrating and blunt trauma can result in injuries to the structures of the portal triad, including the extrahepatic bile ducts, hepatic arteries, and portal vein. Traumatic injuries to these structures are infrequent but are associated with high mortality rates ranging 50-52% for vascular injuries but reported as high as 75% for portal vein injuries.25,26,27 Trauma resulting in extrahepatic bile duct injury carries a mortality rate of 10-31%, and high- grade traumatic injury to the liver has a mortality rate up to 37.5%.28,29 The majority of these mortalities occur on day of admission while in the operating room due to exsanguination from vascular injuries.25,28

The most common causes of penetrating trauma to bile ducts include gunshot wounds, stab wounds, and shrapnel injuries. Penetrating trauma often results in a partial bile duct laceration rather than a complete transection, although transection can be seen in some patients as well.25 Blunt trauma to biliary system is most often secondary to motor vehicle crashes but also includes falls and assaults. (Figure 3) Blunt force to the abdomen can also result in hematoma formation, which can cause extrinsic compression of bile ducts and lead to biliary obstuction.31

Trauma to the liver can result in bile duct injuries. Liver lacerations due to stab wounds or gun shot wounds can result in intrahepatic biliary complications in the form of a biloma or biliary fistula. A recent study of hepatic trauma identified that bile leaks occur more commonly in patients with penetrating trauma, those undergoing operative management, damage control surgery with packing of the liver and those with higher- grade liver injury.30 High-grade liver injuries (grade 3-4 on American Association for the Surgery of Trauma Organ Injury Scale) are associated with bile duct injuries in up to 25% of cases.28 Approximately 2.8-7.4% of patients who experience blunt hepatic traumas of all grades have a biliary complication, most commonly bile leaks with associated biloma formation.28,31

Variant Biliary Anatomy

The existence of aberrant biliary anatomy and anatomic variations of the biliary tree creates an increased risk for iatrogenic bile duct injury. Surgeons may or may not be aware of aberrant biliary anatomy at the time of surgery depending on the type of preoperative imaging obtained.

The duct of Luschka was first described in 1863 as a thin bile duct passing through the gallbladder fossa to join the right hepatic or common hepatic duct, although this can also manifest as a small duct that connects to peripheral right intrahepatic branches.32 The prevalence of the duct of Luschka has been reported in 4.6-30% of dissected liver specimens. The ducts are typically thin, measuring 1-2mm in diameter, with variable sites of origination within the liver and sites of union with the biliary tree.32,33

Magnetic resonance cholangiopancreatography (MRCP) of patients suspected of having pancreaticobiliary disease demonstrated anatomic variations in 24.2% of patients. These anatomic variations included an aberrant right hepatic duct in 4.8%, right posterior hepatic duct in 5.7%, trifurcation in 0.8%, low medial cystic duct insertion in 3.8%, long and short cystic ducts in 1.7% and 0.63% respectively, vascular compression of the common hepatic duct in 2.5%, and 2.3% of patients had more than one anatomic variation.34 Bile duct injuries originating from aberrant biliary anatomy have proven difficult cases for providers that necessitate multiple diagnostic modalities and, in some cases, a multidisciplinary approach, to diagnose and treat the underlying injury associated with the aberrant anatomy.35

DIAGNOSIS OF BILE DUCT INJURIES

Clinical Presentation of Bile Duct Injuries

The most common presenting symptom experienced by patients with bile leaks is abdominal pain, often accompanied by fever, due to the inflammation of the peritoneum caused by bile which may or may not become secondarily infected. Patients may also have abnormal liver function tests at presentation, although patients who have leaks alone often have normal serum bilirubin levels as the bile is pooling within the abdomen.2 An increase in serum bilirubin without elevation of transaminases may represent the presence of a bile duct injury, particularly a bilio-venous fistula where bile can flow down its pressure gradient and into the venous system.36

Biliary strictures often have a more insidious clinical onset with cholestatic symptoms such as jaundice and pruritus. The median time to presentation for patients with biliary strictures is 7 to 9.6 months but ranges as early as 6 days and as late as several years following the inciting surgery.2,37,45 Patients with biliary strictures can present acutely with cholangitis and serious bacterial infections as well.21 Patients may have concomitant biliary strictures and leaks depending on the severity of the injury. (Figure 4)

Diagnostic Imaging Modalities

The biliary tree can be imaged by multiple modalities, including MRCP, hepatobiliary scintigraphy (HIDA scan), and cholangiograms obtained by endoscopic retrograde cholangiopancreatography (ERCP), percutaneous transhepatic cholangiography (PTC), and intraoperative cholangiogram (IOC). Several studies have investigated the efficacy of IOC during laparoscopic cholecystectomy, and overall they have concluded that obtaining an IOC in every cholecystectomy reduces the rate of bile duct injury from 0.42% to 0.21% and increases the rate of intra-operative detection of bile injuries from 21.7% to 76.9%.4,5,6,38,39,40,41 Opponents of universal IOC during cholecystectomy caution against the increased operative time, cost, and limited practical use of the procedure. One large study found that only 82.7% of attempted IOCs were successful, and the estimated additional operating to perform IOC was only 19 minutes.1,39

Bile duct injuries are most often diagnosed from a cholangiogram performed during ERCP or PTC, however many noninvasive imaging techniques, such as MRCP and HIDA scan, can be employed to evaluate the biliary tree as well. MRCP has been utilized to assess the biliary anatomy of donor livers prior to living donor liver transplantations and can detect variant biliary anatomy with a sensitivity of 85.6%.42

HIDA scan is a noninvasive nuclear medicine study used to evaluate the vascular flow and biliary drainage of the liver and extrahepatic bile ducts. The sensitivity of HIDA scan to diagnose bile leaks after laparoscopic cholecystectomy and liver transplantation has been reported to be between 83-100%.43,44 HIDA scan has proven to be successful at diagnosing traumatic bile leaks as well.31 Biliary strictures can also be detected by HIDA scan with a reported sensitivity of 75% and specificity of 97%, often manifesting as a failure of the radiotracer to reach the duodenum after protracted amounts of time.44

Treatment of Bile Duct Injuries

The treatment of bile duct injuries can often require a multidisciplinary approach with several therapeutic possibilities including surgery, endoscopy, and interventional radiology approaches. Bile duct injuries can be managed surgically by re-establishing biliary continuity through reconstructive surgery or by suturing open injuries, although in current practice most biliary injuries are primarily treated endoscopically if at all possible. If patient outcomes are the same, then non- surgical approaches are preferred.45

ERCP for Bile Duct Injuries

ERCP with dilation and stent placement and/or sphincterotomy (as needed) has proven to be a successful method of treating bile leaks and strictures and is the current gold standard and first line treatment for most biliary injuries.12,28,31,36,45-51 These therapies can open up strictures, and allow for stent placement to reduce the pressure gradient across a stricture or the sphincter of Oddi, which directs bile to flow downstream into the proximal duodenum and not out of the leak site. Directing the flow of bile away from the site of injury allows the defect to subsequently heal.

In patients with low-grade bile leaks (defined as bile leaks that are visualized during ERCP only when there is full opacification of the intrahepatic biliary tree with contrast, indicating that more pressure is required to force contrast and bile out of the duct injury site), endoscopic sphincterotomy alone has been reported to lead to successful resolution of the bile leak in 91% of patients, although many patients will receive a stent alone in this situation as stent placement is a lower risk maneuver than biliary sphincterotomy. Patients with high-grade bile leaks (visualized during ERCP before intrahepatic biliary tree opacification, indicating a more brisk flow of bile through the leak site) require either stenting or (rarely) surgical ligation for successful resolution of the leak.46 The reported success rate of treating post-surgical bile leaks with endoscopic placement of biliary stents ranges from 77-100%. 45,46,47,48,49

As one would imagine, less severe bile duct injuries have higher success rates with stenting than more severe or complex injuries. Of the Strasberg classification of bile duct injuries, Class A injuries (leak of cystic duct or small accessory duct) have reported success with ERCP and stent placement in 99% of cases. More severe Class E1 through E5 bile duct injuries (stricture or complete transection of the common bile duct or common hepatic duct) have a reported successful treatment in 77% of attempted ERCPs, but the majority require immediate surgical intervention.48

Bile duct strictures due to surgical injury often require long-term stenting with multiple follow-up ERCPs to obtain resolution of the strictures. Success rates of 71% and 74% have been reported after six and eleven months of stent therapy, respectively.45,47 Anastomotic strictures following OLT and ALDLT are successfully treated by endoscopic dilation and stenting in 74-88.9% of cases.9,12,49 However, up to 18% of patients have stricture recurrence after success initial treatment of the anastomotic stricture, and they may require anywhere from one to four additional ERCPs for successful treatment.13 The simultaneous presence of both a bile duct stricture and a bile leak following surgery is a well-documented risk factor for recurrent strictures.13,45

Endoscopists have long used plastic stents in the treatment of benign, post-operative biliary strictures. Uncovered metal stents, commonly used in the palliative treatment of malignant bile duct strictures, were trialed in 1990s for the treatment of benign post-surgical strictures as an alternative to plastic stents. Although they helped reduce the frequent stent exchanges required with plastic stents, metal stents encountered high rates of stent occlusion and stricture relapse due to ductal mucosal hypertrophy through the stent mesh with resulting recurrent obstruction which proved to be difficult to treat.50,51 Covered self- expandable metal stents (C-SEMS) that are removable have been developed to prevent stent occlusion and may be used for temporary stenting of benign bile duct strictures.52,53,54 Success rates of 88-90% have been reported for C-SEMS in the treatment of benign biliary strictures that were initially refractory to plastic stent treatment. Recurrence rates of 7.1-12% were reported within 16 months of stent removal.52,53 Post-liver transplant anastomotic strictures are often stented using an increasing number of plastic stents to progressively dilate the stricture, although this practice is much less common in the era of fully covered metal biliary stents. Fully covered metal stents, which can provide a similar dilation diameter as several plastic stents, have achieved similar resolution rates in anastomotic strictures.54 Metal stents do carry a risk of migration (ranging 16-24.2%) but most often migrate distally into the duodenum and are expelled without causing clinical complications.53,54 Furthermore, migration is not always to be considered a negative event as it may signify resolution of the underlying stricture.

Percutaneous Transhepatic Catheter (PTC) Drainage for Bile Duct Injuries

PTC is a non-surgical approach that is often attempted after failure of ERCP to treat bile duct injuries. PTC is also performed when bile ducts cannot be accessed by ERCP, often due to post-surgical bowel anatomy (most commonly in patients with a Roux-en-Y gastric bypass).50 PTC is performed by interventional radiologists and provides the same physiologic treatment as ERCP, via dilation and stent placement, but through a percutaneous transhepatic approach. 12,45,55,56 PTC with a bilio-enteric stent placed across the sphincter of Oddi has a reported success rate of 77.2-80% at treating bile leaks.45-50 This success rate is seen with nondilated intrahepatic bile ducts as well, which were once thought to be a contraindication to PTC due to difficult percutaneous accessibility.50 The success rate of treating post-surgical bile duct strictures with PTC dilation and stenting has been reported from 58.8-81%.12,55

Management of Traumatic Bile Duct Injuries

Patients with blunt and penetrating traumatic injuries to the biliary system can often be safely managed non- surgically via ERCP.28,31,36,57,58 Patients with bile leaks secondary to gunshot wounds and stab wounds have had successful resolution of bile leaks when managed non- surgically with sphincterotomy and/or stent placement.57

It is the standard of care to manage patients with blunt hepatic trauma non-surgically.36 One study evaluated non-operative management of severe, grade 4 and grade 5, blunt liver injuries (parenchymal disruption of 25-75% and >75% of a hepatic lobe, respectively) and found that ERCP successfully treated 81% of bile leaks in these cases. Two independent predictors for failed non-operative management were identified: admitting systolic blood pressure less than 100mmHg and presence of additional abdominal organ injury. Failure of non-operative management was seen in 23% of patients with both of these predictors and only 4% with neither.58 Complex bile duct injuries, such as complete transection or stricture of the common bile duct, common hepatic duct, or disruption at the level of hepatic duct bifurcation, almost always require surgical intervention after diagnosis by ERCP.41,48 (Figure 5)

Surgical Reconstruction as Treatment of Bile Duct Injuries

Referral to a tertiary medical center for surgical intervention is often necessary with severe bile duct injuries. Primary end-to-end anastomosis may be performed when the loss of tissue from the bile duct transection is not significant. This technique has been associated with a high rate of stricture formation at the anastomosis site, but nearly two-thirds of these strictures can be successfully managed endoscopically without further surgery.59 When end-to-end anastomosis is not possible, a Roux-en-Y hepaticojejunostomy (RYH) is the preferred treatment. Strictures can occur with RYH as well, but the stricture rate is reduced when the anastomosis is performed more proximally where right and left bile ducts join, rather than at a more distal site in the common hepatic duct.60 Patients with severe bile duct injuries may rarely develop chronic complications such as secondary sclerosing cholangitis with associated cirrhosis or acute liver failure; rarely, liver transplant can be required.61

Morbidity and Quality of Life after Bile Duct Injuries

Bile duct injuries and the subsequent therapies can result in short term and sometimes life long complications for patients.62,63 One study followed patients after surgical reconstruction for bile duct injuries, reporting a mean hospital stay of 17 days following surgery, with 38% requiring acute readmissions within the first year most commonly for cholangitis. At one year follow up, 62% of patients remained asymptomatic, 28% continued to have episodic cholangitis, and 10% had persistently elevated liver function tests but were asymptomatic.62 Another study evaluated patients at a mean of 70 months after treatment of bile duct injury following laparoscopic cholecystectomy, reporting 93% of endoscopic treatments and 84-94% of Roux- en-Y hepaticojejunostomies remained functionally successful without need for additional surgery. Despite the promising reports of these treatment success rates, the patients reported on questionnaires a significantly poorer quality of life on mental and physical scales compared to controls.63

CONCLUSION

Biliary injury can occur via many routes. Laparoscopic cholecystectomy and liver transplantations are the most common surgeries associated with bile duct injuries, but other treatments for hepatic malignancy such as TACE, partial hepatectomy, and radiation may result in bile duct injuries as well. ERCP remains the principal method of diagnosis and treatment using plastic or covered metal stents. Percutaneous and surgical approaches can be utilized as well to properly treat each individual case, and complex bile duct injuries often require multidisciplinary treatment.

Download Tables, Images & References

HEPATITIS C: A NEW ERA OF TREATMENT, SERIES #7

Knowledge Gaps About Hepatitis C Prognosis and Treatment Among Non-Gastroenterologists and Medical Students

Read Article

Given the availability of new Hepatitis C (HCV) therapeutics and increased potential for cure, it is necessary to assess knowledge about treatments among current and future physicians in the United States. Here, the authors aim to quantify the awareness of HCV curability and treatment among gastroenterologists, primary care physicians and medical students.

What is Current Knowledge?

  • New all oral regimens for chronic hepatitis C promise greater sustained viral response (SVR) rates.
  • Underreporting of hepatitis C (HCV) infection has been well documented in the literature.

What is New Here?

  • There is a lack of knowledge about HCV treatments and cure outside of hepatologists and specially trained gastroenterologists.
  • Primary care physicians (PCPs) and medical students are unaware of modern HCV treatments.
  • Experience managing HCV is associated with increased awareness of new therapeutics.

Introduction

In light of Centers for Disease Control and Prevention (CDC) and United States Preventative Task Force (USPTF) screening guidelines for HCV, we aimed to quantify the awareness of HCV curability and management among gastroenterologists, primary care physicians and medical students.

Methods

An online survey was disseminated to several email listservs affiliated with the Tulane University School of Medicine. Four categories of respondents were evaluated with Chi-Squared and Kruskal-Wallis testing: Gastroenterology (GI); Family and Internal Medicine (FIM); Clinical Medical Student (CMS); and Preclinical Medical Student (PMS).

Results

196 responses were analyzed (9 GI, 27 FIM, 90 PMS, and 70 CMS). Analysis identified differences in knowledge of HCV curability (p<0.001), experience managing HCV (p<0.001), and frequency of identifying interferon (p<0.001), ribavirin (p<0.001), or a protease inhibitor (p<0.001) as treatment modalities. GI respondents consistently demonstrated greater knowledge of HCV curability, management, and treatment. Less than 60% of FIM and only 12-30% of medical students were aware that HCV is curable.

Discussion

The FIM, CMS, and PMS groups lacked knowledge concerning HCV treatments and curability, which indicates a possible need for outreach to non-GI specialists and medical training sites.

INTRODUCTION

The hepatitis C virus (HCV) is estimated to infect 1.9-5.2 million Americans, with over 75% of cases becoming chronic.1,2 Many develop hepatic complications, such as cirrhosis or hepatocellular carcinoma, leading to total annual HCV management costs projected to exceed $9 billion by 2024.3,4

In the summer of 2012, the Centers for Disease Control and Prevention (CDC) advocated one time HCV screening for all persons born between 1945 and 1965, which the United States Preventative Task Force (USPSTF) later corroborated.5,6 This recommendation was likely related to the release of first generation direct antiviral agents (DAA), which have drastically increased sustained viral response (SVR) rates.7,8

Though underreporting of both acute and chronic HCV have been confirmed elsewhere,9-13 few studies have examined knowledge among healthcare providers (HCPs) in the United States concerning curability and current pharmacotherapy for HCV. The Institute of Medicine’s (IOM) 2010 report on viral hepatitis acknowledged a gap in provider knowledge about several aspects of HCV management, including the “sequelae of chronic viral hepatitis” as well as in “proper follow-up management for chronic infection”, but did not squarely assess understanding of curability.2 Studies assessing trainees’ knowledge are also scant and could only be identified from Eastern countries.14,15

Given the availability of new HCV therapeutics and increased potential for cure, it is necessary to assess knowledge about treatments among current and future physicians in the United States. We aimed to quantify the awareness of HCV curability and treatment among gastroenterologists, primary care physicians, and medical students.

METHODS

This is a cross-sectional study to assess the knowledge, attitudes and practices (KAP) of HCV pathogenesis and treatment among medical students and physicians at a single academic medical center (Tulane University School of Medicine in New Orleans, Louisiana).

Survey Creation

An online survey was created via GoogleForms, which included demographic (age, sex, zip code) and four study questions, listed in Table 1. Medical specialty or year of medical school education was also collected, as appropriate. To prevent bias by later study questions, Question 1 concerning HCV curability, was presented on a single page before the remaining questions. The word “Cure” was used in the place of “SVR,” as the authors did not believe the idiomatic hepatology language would be equally understood across all specialties. Students and faculty from the single center vetted the survey for clarity before distribution. Complete surveys are available in the online supplemental materials.

Survey Distribution

The KAP survey was initially distributed electronically to listservs of gastroenterology, family medicine, internal medicine and medical students on October 9, 2013. The subject line of the first recruitment email read “Louisiana Hepatitis Study”. Recruitment was considered complete if twenty-five responses were received or the survey was distributed to the listserv three times. When required, surveys were redistributed with the subject line reading: “Quick Survey for Louisiana Hepatitis Study”. A standardized form email was used to recruit survey responses (supplemental materials). Individuals 18 years of age or older who were either a current medical student or medical doctor affiliated with the single center were included.

Statistical Analysis

Data were downloaded from the GoogleSurvey tool, transformed for SPSS and divided into four subject populations for analysis: Gastroenterologist/ Hepatologist (GI), Family and Internal Medicine (FIM), Clinical Medical Student (CMS) and Preclinical Medical Student (PMS). For the study, PMS included first- and second-year, and CMS includes third- and fourth-year medical students.

Chi-squared or Fisher’s Exact were used to evaluate Questions 1, 2, and 4. Shapiro-Wilk testing for normality was performed on the Likert-scale responses of Question 3; none of the four subgroups were normally distributed (GI p=0.037; FIM p<0.001; CMS p<0.001; PMS p<0.001). Kruskal-Wallis testing was then employed for Question 3. SPSS Version 21.0 was used for analysis. Significance was set at a=0.05.

RESULTS

Survey Respondents

A total of 201 survey responses were collected (9 GI, 24 Family Medicine, 8 Internal Medicine, and 160 medical students). The initial survey completion rate among students exceeded expectations, but physician response rate was low and required multiple emails to the GI, Internal Medicine and Family Medicine listservs. Of the 8 Internal Medicine responses, 5 self-identified as specialists other than GI and were not included in the analysis. Family and internal medicine were combined to represent the primary care population. A total of 196 responses were evaluated (9 GI, 27 FIM, 70 CMS, and 90 PMS). The average age and sex distribution by subgroup are presented in Table 2.

Knowledge of HCV Curability (Question 1)

Chi-squared testing showed significant differences when comparing all four groups for knowledge of HCV curability (p<0.001): 100% of GI, 59.3% of FIM, 30.0% of CMS and 12.2% of PMS reported that HCV was curable. Proportional analysis showed that all four groups were significantly different from each other in awareness of curability.

Experience Managing HCV Infection (Question 2)

Chi-squared testing detected a significant difference in experience managing HCV patients (p<0.001). 100% of GI, 81.5% of FIM, 72.9% of CMS and 7.8% of PMS claimed experience managing HCV. Respondents with experience managing HCV had 4.883 the odds of reporting that HCV was curable (95% CI: 2.497 — 9.549; p<0.001).

Perception of Hepatitis C on Patient Health (Question 3)

Kruskal-Wallis testing found significant differences among subject populations on awareness of the extent of injury associated with chronic HCV (p=0.026). Mean and interquartile range was 4 (3-5) for GI, 3.67 (3-4) for FIM, 4.10 (4-5) for CMS and 3.80 (3-4) for PMS. No post-hoc comparisons were significant.

Knowledge of Hepatitis C Treatments (Question 4)

There was a significant difference in proportions reporting interferon as a drug used to treat hepatitis C (p<0.001): 100% of GI, 82.6% of FIM, 78.9% of CMS and 20.0% of PMS cited interferon. Proportional analysis showed that PMS had lower proportional knowledge of interferon than all other groups.

There was a significant difference in proportions reporting ribavirin as a drug used to treat hepatitis C (p<0.001): 87.5% of GI, 56.5% of FIM, 59.6% of CMS and 11.9% of PMS cited ribavarin. Proportional analysis showed that PMS had lower proportional knowledge of ribavirin than all other groups.

There was a significant difference in proportions naming a DAA (telaprevir, boceprevir, or sofosbuvir) as a treatment for HCV(p<0.001): 75.0% of GI, 17.4% of FIM, 14.0% of CMS and 5.0% of PMS cited one or more protease inhibitors. Proportional analysis showed that GI mentioned a protease inhibitor in greater proportions than all other groups.

DISCUSSION

This study exposed shortfalls of knowledge about the prognosis and treatments for hepatitis C. In light of the rapidly evolving field of HCV therapeutics and the recent joint release of guidelines by the American Association for the Study of Liver Disease and the Infectious Disease Society of America on screening and treatment of HCV, it is of great importance to assess healthcare provider knowledge.16

Greater knowledge of curability and treatment among GI was expected, as they are most involved in HCV management and research. With the recent release of second generation DAAs promising fewer side effects, lower rates of complication, and shorter treatment periods,17-19 it may fall on primary care physicians to discuss treatment with the HCV patient. Thus, with only 60% of FIM acknowledging curability for HCV in our cohort, and even fewer capable of mentioning a DAA, there is a clear need for improved outreach.

Similarly, underreporting of HCV is a serious issue increasingly documented in the literature.12,13,20 A recent NHANES report showed that only 32-38% of all HCV antibody positive people in the United States received follow-up care.12 Specialist referral of diagnosed HCV patients has failed to surpass 50% in several cross sectional studies.21,22 Definitive reasons for undertreatment and underreporting are not answered by the study, but it is reasonable to assume insufficient incentive for either with limited knowledge regarding its curability.

The low level of insight among medical students regarding HCV curability and pharmacological agents suggests that a push towards graduate medical education is necessary. Interestingly, CMS were able to mention interferon and ribavirin at higher rates than PMS, suggesting some exposure to early HCV treatments in clinical years.

There were several limitations to this study. Firstly, the data set was restricted to a single academic center, limiting generalizability and may be skewed by the university’s educational programming. Additionally, the initial protocol compared subspecialists with primary care and medical students, but the survey response rate among physicians was lacking and consequently subspecialists were excluded from analysis. Although respondents were instructed not to use outside resources, the online format for the survey created potential for bias by respondents researching answers.

CONCLUSION


We identified a gap in knowledge about HCV treatments and curability among primary care providers and medical students. Though a full policy discussion is beyond the scope of this paper, we recognize the need for a national assessment and possibly improved dissemination of information concerning HCV treatments to non-GI specialists and medical trainees. Without an appropriate fund of knowledge amongst medical trainees and general practitioners, patients with chronic HCV are wont to face difficulty in obtaining appropriate medical referrals for the most up-to-date treatments.

Download Tables, Images & References

GASTROINTESTINAL MOTILITY AND FUNCTIONAL BOWEL DISORDERS, SERIES #5

Cyclic Vomiting Syndrome: Diagnostic Criteria and Insights into Long Term Treatment Outcomes

Read Article

The alternating pattern of disease and disease-free periods distinguishes cyclic vomiting syndrome (CVS) from other disorders of nausea and vomiting. This entity has been increasingly recognized in adults and has resulted in significant morbidity and poor quality of life. Recent referral patterns suggest prevalence of up to 0.2% in the adult population and an explanation for nausea and vomiting in 12% of a referral population to a teaching academic center. In this article we discuss diagnostic criteria and insights into long term treatment outcomes.

Chad J. Cooper MD, MHA Richard W. McCallum, MD, Professor and Founding Chair, Department of Internal Medicine Division of Gastroenterology Texas Tech University Health Sciences Center, El Paso, TX

INTRODUCTION

Cyclic vomiting syndrome (CVS) in adults is a disorder characterized by recurrent abrupt bouts of nausea, vomiting and abdominal pain separated by variable periods of normal health. This alternating pattern of disease and disease-free periods distinguishes CVS from other disorders of nausea and vomiting. This entity has been increasingly recognized in adults and has resulted in significant morbidity and poor quality of life.1 Recent referral patterns suggest prevalence of up to 0.2% in the adult population and an explanation for nausea and vomiting in 12% of a referral population to a teaching academic center.2

Diagnostic Approach

CVS is really a diagnosis of eliciting a “classic” history of this disease. Patients typically present with a variable number of episodes of nausea, vomiting and mid-epigastric abdominal pain per year. During all presentations in adults there is accompanying severe mid-epigastric abdominal pain, with or soon after the onset of nausea and vomiting. This history of mid- epigastric abdominal pain tends to attract the need to exclude other sources since it can mimic an acute abdomen. Diagnostic Criteria of CVS is based on Rome III which includes the following list in (Table 1). These criteria were developed when CVS was predominantly only being recognized in children and abdominal pain was not a predictable concomitant feature.

The majority of CVS attacks occur without any warning although in retrospect patients report that up to 60-80% of CVS attacks can be associated with a trigger mechanism such as infection (chronic sinusitis and upper respiratory infections), psychological stress, emotional stress, physical stress (heavy exercise), lack of sleep, diet (chocolate, cheese), motion sickness and onset of menses.3-4 Many patients take hot showers or baths during the vomiting episodes and report a decrease in symptoms and therefore the contact with hot water is assumed to have a “relaxation effect”.

As far as contributing factors or etiologic “subgroups” approximately 24-70% of CVS patients report a personal or family history of migraines.1 However in adults the migraine subset is more in the range of 30-40% of the CVS population. Psychiatric disorders, such as anxiety, depression are frequent comorbid findings in CVS patients.6 The anxiety present in CVS patients, including panic disorder, has been reported to trigger attacks in 66% of cases.7 Anxiety can also be increased as a result of the burden of the illness, anticipation of the next vomiting episode or psychological trauma and experiences prior to the onset of CVS.3,7 Sometimes psychological disorders in CVS patients including depression are so dominant that co- management with a psychiatrist may be indicated.

Another identifiable subgroup is diabetes mellitus which is increased to 15% in the CVS populations compared to approximately 8% in normal population based studies. The theory proposed here is that elevated glucose levels, usually early in the course of diabetes, in the setting of genetically predisposed CNS chemoreceptors can trigger a vomiting cycle.

On review of the current literature of cyclic vomiting syndrome in adults marijuana use is present in 42-53% .8 The predictable scenario is a typical pattern of daily marijuana intake beginning in the teenage years for recreational use. The cyclic episodes of vomiting do not occur until at least 5 years of chronic daily use. A clinical entity termed cannabinoid hyperemesis syndrome has also been separately described and is actually the same clinical presentation as CVS.8 Cannabinoid hyperemesis syndrome is characterized by chronic marijuana use, cyclic episodes of nausea, vomiting, abdominal pain and frequent relief with taking a hot bath.9

Not all cannabis users develop cannabinoid hyperemesis syndrome. However the cumulative dose of marijuana, genetic factors, and psychological parameters may contribute to this condition. The pathophysiology of cannabinoid hyperemesis syndrome is unknown. There are some hypotheses proposed to explain this phenomenon: (a) accumulation of cannabis derivatives in the brain based on their lipid solubility and long-term half-life, (b) degradation of the cannabis ingredients to some potential emetic metabolites or toxins, (c) delayed gastric emptying induced by cannabis and (d) down-regulation or desensitization of the cannabinoid receptors due to chronic cannabis use.8,10 Another theory is that chronic cannabis is associated with inhaling toxins related to the various sources and preparation of the marijuana and over time these toxins could accumulate in the CNS.11 However, the major message is chronicity of >5 years with daily use leading to increasing storage in fat tissue in the brain and in those genetically susceptible individuals nausea receptors are activated. This setting is to be distinguished from the “legal use” of marijuana for cancer and pain related indications. In addition cannabis has been shown to acutely delay gastric emptying and this can also contribute to inducing a vomiting cycle.9,11

In our experience at an academic gastrointestinal motility referral center, we recently reviewed a total of 48 patients diagnosed with CVS, 37 females and 11 males with a mean age of 34.8 year old. The majority of the patients reported cyclic episodes occurring approximately every 2-3 months. Five (10%) patients had relief of symptoms with hot baths or showers and 11 (23%) had worsening of symptoms with stress, menses or sleep deprivation. Comorbidities included diabetes mellitus (31%), hypertension (23%), hyperlipidemia (15%), anxiety (48%), depression (25%), migraines (40%), family history of headaches/migraines (31%), panic disorder (11%), and chronic daily marijuana use (23%) for more than 5 years. Eleven (23%) patients were smokers, 7 (15%) had a history of alcohol use and 15 (33.3%) were given narcotics acutely at some time throughout the course of their disease during ED visits. Six (20%) of patients reported a significant disruption in their professional and/or personal social life. Fifteen (19%) patients had a cholecystectomy. Twenty-five (52%) had frequent ED visits before being diagnosed and treated.

Adult patients typically have been symptomatic for a long time before diagnosis. Patients often remain undiagnosed for some time due to lack of recognition of this clinical entity with reports suggesting a delay in diagnosis for as long as 8-21 years following onset of the symptoms.6,12 Over time without appropriate specific treatment, CVS cycles slowly begin to “coalesce” and become closer together and this can confuse the presentation and suggest more of a “chronic” entity such as gastroparesis. CVS results in a significant morbidity for patients with loss of time at work or school, a significant disruption in professional and personal life as well as an economic burden.9,12-13 CVS in adults can range from mild disease with infrequent episodes to severe debilitating disease requiring multiple emergency department (ED) visits and frequent hospitalizations.12 These patients often undergo multiple unnecessary diagnostic tests and procedures without any apparent clinical benefit. Abdominal pain and mild leukocytosis have prompted unnecessary cholecystectomies and other abdominal surgeries because CVS was mistaken for an acute abdomen.

Gastric Emptying Studies in CVS

Even though CVS has been increasingly recognized in the adult population, there is a lack of data as to the gastric emptying (GE) pattern. Using a standardized 4 hour egg beater scintigraphic method a normal GE is defined as < 90% retention at 1 h, < 60% at 2 h, and < 10% at 4 h. Rapid GE is defined as < 35% isotope retention at 1st hour and/or < 20% at 2nd hour.14 Delayed gastric retention is defined as a delay of greater than 90% at 1 h, 60% at 2 h, and 10% at 4 h based on normal data established for this standardized GES.14 Employing these criteria for rapid GES we found that 30% met these criteria while 70% had a normal GES. The GE test was performed during the remission phase of CVS. Delayed gastric emptying was not identified. Our group also published criteria for rapid gastric emptying as being <50% isotope retention at 1 hour and the majority of our adult patients (65%) with CVS had a rapid GE and 35% had a normal GE.15

A rapid or normal GE can therefore be used as confirmatory evidence of CVS so that clinicians can confidently exclude gastroparesis from the differential. Gastric emptying studies should be performed during the remission phase when there are minimal or no symptoms and no narcotic medications are being received.16-17 Gastric emptying studies while patients are in the hospital receiving narcotics are discouraged. Narcotics inhibit GE thus producing a slow gastric emptying result leading to a mislabeling of these patients as having gastroparesis. In addition, the preceding use of marijuana can delay GE. Explaining the role of a rapid GE during a vomiting free period has led to speculation and support for an underlying autonomic dysfunction is these patients as well as evidence for increased serum ghrelin as another factor in speeding up the GE.11

Diagnostic Evaluation

The diagnosis of CVS requires that other known and treatable disorders be excluded. The differential diagnosis for patients with CVS that should be ruled out includes those listed in Table 2. We recommended a diagnostic algorithm (Figure 1) through which a patient presenting with an acute episode of nausea, vomiting, epigastric abdominal pain should be evaluated so that other diagnoses can be excluded by history, physical examination, and basic laboratory studies including a complete blood count (CBC), complete metabolic panel (CMP) with liver function tests, amylase, and lipase, a urinalysis as well as an upper GI series/small bowel follow through.6 An abdominal ultrasound may help in evaluation of possible gallstones, pancreatitis and ureteropelvic junction obstruction. An esophagogastroduodenoscopy (EGD) should be performed in patients with acute vomiting with or without hematemesis to exclude gastric outlet obstruction or peptic ulcer disease as well as H. pylori. Imaging studies such as an abdominal CT should be considered to exclude structural lesions.

The decision as to which diagnostic tests to perform should be tailored to the clinical presentation of the patient. In adults, much consideration must be used to differentiate CVS from gastroparesis. A subset of patients with idiopathic or diabetic gastroparesis presents with cyclic emetic episodes similar to CVS.7 Patients with gastroparesis exhibit more chronic daily symptom severity and a delayed gastric emptying on scintigraphic study. In contrast, gastric emptying is often accelerated or normal and not delayed in patients with CVS during the asymptomatic period when vomiting is absent.7

Management

Once a cyclic vomiting episode is in progress, supportive measures are at the forefront of management. Intravenous fluids should be given to prevent dehydration and electrolyte imbalance. The approach to treatment in the ED setting is inducing sedation, sleep and relaxation mainly through IV lorazepam (1-2mg every 4 hours) with support from narcotics, anti-histamines and antiemetics to terminate the emetic phase although hospitalization is often required to achieve this goal. Family involvement is a crucial part of management in order to cope with an unpredictable, disruptive, unexplained illness that is commonly misdiagnosed.

Long term treatment of CVS is based on trying to identify the etiologic subgroups particularly the role of psychological stress while prescribing prophylactic drug and abortive therapy and supportive measures to ameliorate acute vomiting episodes. In relation to psychological stressors, stress management techniques as well as daily lorazepam (1mg up to every 6 hours) will help relieve anxiety. For the subset of patients with significant depression, co-management with a psychiatrist may be indicated to select antidepressant therapies with the least likelihood of exacerbating the emetic illness. It is appropriate to start anti-migraine prophylaxis in those CVS patients with a positive family history or personal migraine history. Anti-migraine drugs that are effective at reducing the number of episodes or severity of migraines include sumatriptan, propranolol and topamax. Patients with a history of chronic cannabinoid use should be counseled in regards to cessation that commonly leads to symptomatic improvement.4,11 Studies from our patient population have shown a high rate of cannabis use in a subset of patients with CVS. These patients need a higher dose of amitriptyline for the control of their CVS attacks compared to non-cannabis users.18-19 Therefore, it is important to identify cannabinoid hyperemesis syndrome as part of CVS, since a long term goal is decreasing and stopping cannabis use in these patients.

Long term management is focused on reducing and actually preventing future hyperemesis episodes .20-21 At the forefront of CVS management, tricyclic antidepressants (TCA), especially amitriptyline, have been shown to be effective for pharmacological prophylaxis. They are well tolerated and very effective in treating adult patients with CVS in doses of 50 to 200mg as necessary and as tolerated. Tricyclic medications act by decreasing cholinergic neurotransmission and modulating alpha-2-adrenoreceptors, thereby reducing the sympathetic nervous system and brain— gut autonomic dysfunction.13,22-23

The treatment approach with tricyclic antidepressants requires beginning with a low initial dose of amitriptyline 10 mg at night with incremental increases in 10 mg doses every 2 to 4 weeks to titrate to the desired therapeutic effect.24-25 There is no established dose to control the symptoms but prevention of the vomiting cycles is the goal. Side effects of using TCAs include dry mouth, somnolence, constipation, postural hypotension, chronic fatigue, blurred vision and mild hallucinations.24-25 Side effects can be minimized by slowly increasing the dose by 10 mg every 2—4 weeks. The rational for this approach is to identify what is the lowest dose that may be therapeutic in an individual and still limit side effects that can occur with higher doses. Tricyclic antidepressants take more than 1 month to achieve full therapeutic effect following initiation and this must be conveyed to the patient. Other TCAs such as nortriptyline and doxepin can be used as substitutes with less adverse events, but still with therapeutic benefits. More recently, the anticonvulsant agents zonisamide (100-600mg daily) and levetiracetam (500-1000 twice daily) agents have demonstrated efficacy in adult patients who are unresponsive or intolerant of TCAs, but their current role can only be considered as second line therapy.7,26

Once initial control is achieved with escalating amitriptyline dosing and concurrent lorazepam for anxiety, supportive therapy involves antiemetic agents include ondansetron, promethazine or prochlorperazine for breakthrough nausea. The antispasmodic (dicyclomine) is for irritable bowel syndrome (IBS) like abdominal pain, especially in patients with rapid gastric emptying and an exaggerated gastro-colic reflex. Proton pump inhibitors can be briefly used for gastroesophageal reflux symptoms in relation to excessive vomiting.

In our experience at a gastrointestinal motility referral center, we found that 83.3% of our patients who began on a low dose (10mg) of amitriptyline before bedtime were able to gradually escalate using an approach of 10mg increments every 2-4 weeks as tolerated and achieve symptom control as evidenced by preventing relapses and ED visits. This titration approach to the dosing of amitriptyline achieved symptomatic relief in 8% patients at a dose of 50- 75mg; 50% at 100mg; 21% at 150mg and 8% at 200mg. Nonresponse to standard therapy in adult cyclic vomiting syndrome patients occurs in approximately 13% and is not explained by under dosing with TCA therapy.24 The main risk factors for nonresponse to amitriptyline are: co-existing poorly controlled migraine headaches, psychiatric disorder, chronic narcotic and ongoing marijuana use, which should be addressed aggressively when symptom exacerbations continue during attempts to induce remission in cyclic vomiting syndrome with high-dose TCA therapy.24

Long term outcomes are now becoming apparent as treatment patterns become recognized. One pattern requires further increasing the maintenance dose overtime due to some dose tolerance slowly occurring with breakthrough vomiting cycles. Another group in whom symptoms were controlled for at least one year with no ED visits were able to be successfully tapered to a lower dose with no ED visits. The amitriptyline dose is slowly tapered or even stopped over time, usually at least 1 year. Twenty-one percent of our patients were able to reduce their dose to 10-20mg per day. One additional incentive we witnessed was a pregnancy goal in females since amitriptyline is listed as a category C by the FDA and therefore is not recommended during pregnancy.

In Summary, treatment with TCA is an effective strategy in 87% of patients and significantly decreases the frequency of attacks, number of emergency room visits and hospitalizations. Once symptoms are controlled for at least 12 months the dose of TCA can slowly be tapered to reach very low doses or even be stopped while maintaining symptomatic control. This has been a new observation, namely that effective tapering over 6 to 12 months can be achieved. The theory to explain this observation is that the CNS receptor hypersensitivity initially present in CVS patients has been successfully blocked during treatment with amitriptyline providing a time frame where the recognized risk factors of migraine, stress, diabetes, marijuana can be addressed and better controlled. Hence, the environment affecting CNS nausea receptor sensitivity has now changed so that the protective role of amitriptyline is no longer required.

CONCLUSIONS

The intent of this article was to further characterize the clinical presentation and propose new diagnostic criteria for CVS in the adult population (Table 3). CVS is not a rare condition in adults as it was once thought to be and is essentially a bedside diagnoses based on the “classic” stereotypical cycles of the vomiting episodes and discrete symptom-free intervals. It is now more common in adults than children and is diagnosed in up to 12% of patients being referred for evaluation of nausea and vomiting. As cyclic vomiting episodes coalesce, usually because no specific therapy has been given, patient symptoms can begin to resemble gastroparesis, a disorder presenting with more continuous chronic nausea and vomiting but typically much less abdominal pain. However, subsets of diabetic gastroparesis patients can have relapsing “cycles” superimposed on this chronicity and this is on example where the distinction from CVS becomes very difficult. Overall the epigastric abdominal pain is a more dominant complaint among CVS patients which is not the usual case in gastroparesis. However, rapid or normal gastric emptying in CVS may be a final “tie-breaker” in distinguishing it from the slow gastric emptying of gastroparesis.

The important message for our clinicians is that CVS can be presented to the patient as a potentially reversible disease: Following initial intensive treatment to achieve remission and after effectively addressing comorbidities, the dose of amitriptyline can be slowly tapered and even stopped over time.

Take Home Message

CVS in adults is under diagnosed and improved awareness and recognition of disease characteristics can help reduce invasive and costly diagnostic workups that have a negative effect on patient economics. This is particularly so for physicians in the Emergency Department who should have a high clinical index when a pattern of unexplained episodes of acute nausea, vomiting and abdominal pain, are observed with frequent ED visits and hospitalizations. This is the CVS “blue print” and epigastric abdominal pain should be added to the Rome criteria for adult CVS patients. Further key clinical clues for CVS lie in appreciating the comorbidities of anxiety, depression, migraine headaches and diabetics. Chronic marijuana has now become a new and very prevalent etiology and should be added as supplementary criteria. Finally, CVS patients have either a rapid or normal GE. This is a “signature” finding separating CVS from gastroparesis and we strongly recommend that GE status be added as one of the major criteria for the diagnosis of CVS.

List of Abbreviations: CVS: Cyclic vomiting syndrome; GES: Gastric emptying study; GE: Gastric emptying; ED: Emergency department; TCA: tricyclic antidepressants; PPI: proton pump inhibitors

Download Tables, Images & References

EPIDEMIOLOGY OF GASTROINTESTINAL CANCERS, #5

Dissecting the Epidemiology of Pancreatic Adenocarcinoma

Read Article

Here we discuss the epidememiology of Pancreatic cancer (PCa), the eighth leading cause of death from cancer in men and the ninth leading cause of death from cancer in women throughout the world. Delay in diagnosis, surgically inaccessible location of the pancreas, absence of classic symptoms of the disease and poverty of molecular biomarkers result in a diagnostic challenge.

1Janeesh Sekkath Veedu, MD, 1Febin John, MD, 2CS Pitchumoni, MD, 1St. Peter’s University Hospital Rutgers RWJ Medical School, New Brunswick, NJ 2Chief of Gastroenterolgy, Hepatology and Clinical Nutrition, Saint Peter’s University Hospital, New Brunswick, NJ

INTRODUCTION

Pancreatic cancer (PCa) is the eighth leading cause of death from cancer in men and the ninth leading cause of death from cancer in women throughout the world.1 The American Cancer Society estimated that 45,220 Americans will be diagnosed with Pancreatic Cancer in 2013. Delay in diagnosis, surgically inaccessible location of the pancreas, absence of classic symptoms of the disease and poverty of molecular biomarkers result in a diagnostic challenge. PCa represents 2.4% of all cancers and 3.7% of cancer deaths. Early surgical resection has a survival benefit, however, since the disease is often diagnosed at late stages surgery is not curative and adjuvant therapy becomes palliative.2

PCa is usually seen in the elderly with a male predominance, the peak incidence being in those aged 65-75 years. Adenocarcinoma accounts for 95% of all cases, about 85% are sporadic with no family history or predisposing genetic syndromes. Although 5-year survival is low (<5%), high volume surgical centers have reported survival rates of up to 40%.3 Recent studies have shown that distinct molecular subtypes of PDAC exist and are associated with different prognosis and therapy response.4 Related to improved life-expectancy and probably adoption of cancer associated lifestyles the incidence is growing globally.3

ETIOLOGICAL ASSOCIATIONS

A. ENVIRONMENTAL

Few modifiable risk factors have been implicated in the etiology of PCa.

Cigarette Smoking

Cigarette smoking is a well-established risk factor for PCa and a co-factor in chronic pancreatitis secondary to alcoholism.5 It is attributable for about 20-30% of cases of PCa.6-13 Studies report a higher risk among current smokers compared to non-smokers, up to 6 fold depending on duration and intensity of cigarette smoking,14 (RR=1.74, 1.61—1.87) and also in former smokers with respect to never smokers (OR=1.20, 1.11-1.29).7,9 Smoking 1 pack/day increased the risk by 1% and the risk doubled for those with >40 pack years of smoking.7 Pipe/cigar smoking had lower risk when compared to cigarette smoking but passive smoking (workplace/household) did not increase the risk.12 Though the risk remained elevated for up to 15 years after quitting, a non-significant drop in risk was observed after 20 years.7,9

Cigarette smoke contains nearly 4000 chemicals of which more than 60 has been identified as carcinogens14 (polycyclic aromatic hydrocarbons, N-nitrosamines, aromatic amines, 1,3-butadiene, benzene, aldehydes and ethylene oxide). The toxins reach the pancreas indirectly via bloodstream or biliary regurgitation to exert carcinogenic effect.8,11,15 Recent studies have looked into genetic variations at carcinogen-metabolizing enzymes to further understand individual susceptibility to PCa.15 Of the carcinogens the most potent metabolite, NNK mediated pathways is well studied.14 There have been no major studies on the effect of e-cigarettes on PCa. E-cigarette users were more nicotine dependent than nonusers, had more prior quit attempts, and were more likely to be diagnosed with thoracic and head or neck cancers.16

Alcohol

An association between alcohol abuse and pancreatic injury was reported by Friedreich as early as 1878.17 Freidreich recognized an association of alcohol abuse with chronic pancreatic injury.17 Several studies have evaluated the association of alcohol and PCa but conclusive evidence is lacking.11,17-21 This could be due to interplay of significant confounders such as smoking, pancreatitis, nutritional and genetic factors.17,21 However, alcohol has been projected as an independent risk factor, attributable to 2-5% of all PCa cases (where population prevalence of heavy drinking is 10-15%).11,17,21 Heavy drinking (>40g or >3drinks/ day) is associated with moderate risk (RR=1.22, 1.12- 1.34) in women and up-to 3.5-fold risk in male binge drinkers (>70g or >5 drinks/day).17,19,21 The risk with type of beverage consumed (wine, beer, liquor/spirit) is variable but an increased risk with the duration of alcohol consumption is reported.18,19

The causal role of alcoholic pancreatitis which is responsible for <5% of PCa cases is not adequate enough to explain the link between alcohol and this PCa.21 Acetaldehyde (oxidative pathway) and fatty acid ethyl esters (non-oxidative pathway), the metabolic products of alcohol, activate pancreatic stellate cells leading to inflammation, immune response and cancer.17,21,22 Folate depletion leading to defective DNA synthesis/repair and carcinogen activation via induction of cytochrome P450 is also postulated means of alcohol injury.18

Diet

The role of diet in the pathogenesis of PCa is weak and contradictory. Mediterranean diet rich in plant-based foods, whole grains and fish with modest consumption of meat and dairy products was associated with a decreased risk (OR=0.51, 0.31-0.84).23-26 A 2.4-fold risk was reported in men on a Westernized diet (red/processed meat, potato, sugary beverage, refined grains, eggs and high-fat dairy).24 Red meat consumption was associated with an increased risk in men.27 A statistically significant 19% higher risk in those consuming processed meat (50g/day) was reported in a meta-analysis.27 Energy- dense diet consumption escalated the risk (up to 72%) while soft drinks did not.28,29

Although dietary fat is not associated with increased risk, a recent study attributed a diet rich in cholesterol with low fiber and folate to the increased incidence of PCa in Poland.30,31 Dietary magnesium especially in overweight men is found to decrease the risk (18% reduction with 100mg increased intake) and cruciferous vegetables (OR=0.90) were protective.32,33 There is no protective effect for antioxidant consumption while a 2-fold risk is seen in those with high serum levels of 25 (OH) vitamin D (=100 nmol/L).34,35 Based on data from the European Prospective Investigation into Nutrition and Cancer Cohort, coffee (total or decaffeinated) and tea consumption are not related to the risk of PCa.36

Occupation

Certain occupations are associated with an increased risk, especially with exposure to chlorinated hydrocarbons and polycyclic aromatic hydrocarbons (PAH).37 Dry-cleaning, metal-related work (Gold/silver smith) and electronic work have exposure to chlorinated compounds.37 Although an increased risk with herbicide and fungicide (not insecticide) was noted in one US study, others failed to demonstrate this risk.37,38 PAH associated risk was seen among metal workers and those in aluminum industry.37,39 Occupational exposure to inorganic dust, asbestos and ionizing radiation also amplified the risk.37,40 Assessment of PCa risk among night shift workers in Japan and food industry workers in Finland did not yield significant result except in Finnish males (SIR=1.5, 1.13-1.96).41,42

Exposure To Heavy Metals

There are reports in literature to support the risk for PCa from exposure to heavy metals. Higher incidence of PCa in the East Nile delta region of Egypt is now attributed to cadmium exposure from fertilizers and polluted river water.43,44 An epidemiological study from Louisiana reported this heavy metal exposure from food (pork, seafood, rice) as the cause of increased incidence of this cancer among the Cajuns.45 Cigarette smoke is another potent source of cadmium which could be implicated for the increased risk of PCa among smokers.45 Cadmium exerts its carcinogenic effect via impairment of DNA repair mechanisms to cause genomic instability.44

Arsenic with similar carcinogenic mechanism is also incriminated in PCa.44 A recent study from Florida reported a significant increase in the risk among those living within 1 mile radius of Arsenic-contaminated wells.46 There are reports about childhood arsenic exposure (from milk powder) and increased mortality associated with this cancer.44,47 Although asbestos exposure from drinking water was reported to increase PCa, subsequent follow-up and analysis failed to prove this.48 Similarly there are studies, which link exposure to lead and decreased levels of selenium (toenail concentrations) to PCa.44

Radiation

In a study of the Hiroshima and Nagasaki, the two sites of atomic bombings, no radiation effect was noted.49 However one study showed excess deaths from PCa in patients who received therapeutic irradiation for ankylosing spondylitis50 and another reported two cases of PCa in patients who got abdominal radiation for testicular cancer.51

Infection

Studies have revealed an infective etiology of PCa, the main agents being H pylori, HBV, HCV and HIV.

H pylori, extensively studied as a gastric carcinogen is being investigated for extra-gastric associations. Several studies including a meta-analysis have found significant association of PCa (AOR=1.38, 1.08- 1.75) with this highly prevalent infection (40% in developed countries and 70% in developing countries) and about 2-fold risk with CagA +/ VacA + strains.52,53 The antral colonization of H pylori and subsequent hyperchlorhydria leading to increased pancreatic secretions and hyperplasia is one of the plausible mechanisms.52,54 Inflammation (IL 8 and VEGF) and bacterial overgrowth from hypochlorhydria (increased N nitrosamine) are other suggested mechanisms.52-54

An increased risk was seen with active (RR=3.83), chronic (RR=1.39) and past (RR=1.41,1.06—1.87) HBV infection.55 There was a synergic increase in the risk in chronic/inactive HBsAg carriers with DM.56 HCV with mechanisms similar to HBV was found to double the risk for PCa (SIR=2.1, 1.4, 2.9).57-59

There is an increased incidence of PCa in HIV patients (SIR=2.2, 1.2-3.6).60 PCa was diagnosed at a younger age with advanced stages at presentation and had a higher likelihood of unfavorable performance status in HIV positive subjects.61 Periodontal disease, and Porphyromonas gingivalis, a pathogen for periodontal disease, are reported associations in PCa.62

B. CHRONIC PANCREATITIS

All types of chronic pancreatitis predispose to PCa (RR=5.1, 3.5-7.3), although <10% is attributed to it.63,64 A significant risk was associated with both acute (HR=9.1, 3.81-21.76) and chronic pancreatitis (RR=13.1, 6.1-28.9).63,65,66 Different forms of chronic pancreatitis such as hereditary, autoimmune and tropical pancreatitis are discussed in literature, all of which are significantly associated with PCa. Patients who underwent surgery for the treatment of chronic pancreatitis had significantly lower incidences of pancreatic cancer. Surgery for chronic pancreatitis may inhibit the development of pancreatic cancer in patients with chronic pancreatitis.67 Acute pancreatitis may be an initial manifestation of PCa.

1. Hereditary Pancreatitis (HP)

Hereditary Pancreatitis (HP) is an inherited form of chronic pancreatitis characterized by recurrent episodes of pancreatitis since childhood.68-71 Mutation in the cationic trypsinogen gene (PRSS1) was the first identified genetic defect.68,71 Subsequently several germline mutations such as protease serine 2 (PRSS2), pancreatic secretory trypsin inhibitor (SPINK1), CFTR, chymotrypsinogen C (CTRC) and calcium-sensing receptor (CASR) were discovered.68 Individuals with HP have a high risk for PCa (SIR=87, 42-114).69,70,72 The cumulative risk by age of 75 years is about 40%- 53.5%.69,70,72 Smoking and diabetes further increased the risk in these patients.71

2.Tropical Calcific Pancreatitis (TCP)

Also known as fibrocalculous pancreatic diabetes (FCPD), Tropical Calcific Pancreatitis (TCP) is a form of chronic pancreatitis in Afro-Asian countries.73,74

The exact etiology for this form of chronic pancreatitis has not been established. Studies have clearly shown that this is a high risk factor for PCa (RR=5, 1.03-14.6).73,75 Patients who develop PCa are younger compared to the denovo form. The entity although found in many states in India is well studied in large series of patients mostly from states of Kerala and Tamil Nadu.

Early studies identified TCP as a disease in young malnourished individuals with poor prognosis leading to diabetes and having a high risk for PCa. But recent research found strong genetic links to this disease (SPINK1/CFTR mutations) and dismissed the notion of regional predominance, links to nutrition and grave prognosis.76,77

The pathogenesis of malignancy in pancreatitis is postulated via inflammatory mediators, activation of signaling pathways (cyclooxygenase2 expression, Notch signaling, Hedgehog signaling) and oxidative damage.66,67 Ueda et al. reported a decreased risk in chronic pancreatitis patients managed surgically (HR=0.11, 0.014-0.80) which provides further evidence for the inflammatory etiology.67

3. Autoimmune Pancreatitis

Autoimmune Pancreatitis is a steroid responsive type of chronic pancreatitis, which mimics PCa.78-80 Although several case reports have been published, conclusive evidence regarding its association with cancer is lacking other than an increased occurrence of K-ras mutations.78-80 AIP features a significant inflammatory phase, and hence it is biologically plausible that AIP patients are similarly at increased risk for developing PCa. The potential for systemic inflammation in this multiorgan disease could also contribute to risk for extra pancreatic cancers. Finally, the late age at presentation of type 1 AIP and reports of cancer being discovered shortly before and after AIP diagnosis have fueled speculation that AIP is a paraneoplastic manifestation of an underlying cancer.81

C. DIABETES AND PANCREATIC CANCER

This is a clinically important but controversial topic. Type 2 Diabetes (DM) with its temporal association with PCa is described both as cause and result of the cancer.82-84 This is an independent risk factor with approximately two-fold increased risk compared to general population.82,83 Risk is inversely associated with the duration of DM, the highest risk being with <1year of DM (OR=5.38;3.49—8.30).83 No increased risk was seen in subjects with >9 years of DM (OR=1.02; 0.68- 1.52) which contradicted the findings of a previous meta-analyses.82,85,86 The association between DM and PCa was not modified by gender, smoking, age, or BMI.82 History of diabetes in a first degree relative increased the risk (OR=1.37, 1.10-1.71) per the PACIFIC study (pancreatic cancer: investigation into finding causes).87 A meta-analysis observed equal risk in diabetic men and women but some disparity exists in this regard.83,84,88 Higher risk was seen among those using insulin compared to those without (OR=3.34 vs. 1.50) in the Iowa Women’s Health Study (IWHS).84

Hyperglycemia associated with altered glucose metabolism, chronic inflammation, oxidative stress, and activation of insulin signaling cascades increases the risk of pancreatic cancer.89 The development of DM within a few years of a pancreatic cancer diagnosis is more likely to suggest an effect of the tumor, whereas diabetes of longer duration is more likely to contribute to the development of cancer.90

However in a study from Japan, PCa was diagnosed within 2 years of DM onset (new-onset) in 0% of the patients with early-onset DM, and in 33% of those with late-onset DM. Pre-existing type 2 diabetes, acute alcoholic hepatitis, acute pancreatitis, cholecystitis, and gastric ulcer independently or jointly predict subsequent pancreatic cancer risk.91

The notion that new-onset diabetes in pancreatic cancer is a paraneoplastic phenomenon caused by tumor secreted products was strengthened by a recent study that proposed adrenomedullin, a 52 amino- acid polypeptide, as a strong candidate for mediator of diabetes in pancreatic cancer. Adrenomedullin was also shown to be overexpressed in human pancreatic cancer and plasma levels of adrenomedullin were also increased in pancreatic cancer patients, especially those with diabetes.92 Earlier concept of beta cell destruction has given way to the role of hormonal secretions from the tumor causing insulin resistance, up-regulation of IGF- 1 leading to carcinogenesis via enhanced angiogenesis and cell growth without apoptosis.82,83,89 Supported by the fact is the observation that IGF receptor and insulin receptor substrate-2 (IRS-2) are over-expressed in cancer cells of the pancreas.93 Other+ studies have shown the presence of diabetogenic factors (2030 MW peptide, Amylin/IAPP) in the serum.83,94 Thus patients with new-onset DM with a family history of DM should be screened for underlying malignancy.87 Similarly new-onset DM in older patients (>65 years) with a negative family history and low BMI (<25) or recent weight loss (>2kg) also have a likelihood for associated PCa.95 Reducing diabetes by controlling obesity could benefit pancreatic cancer rates, in addition to the many other known health benefits.82 One study showed that dyslipidemia, but not diabetes, is a significant risk factor for PCa. Patients with new-onset diabetes and a history of dyslipidemia are at an especially high risk of PCa.96 DM is also an independent risk factor for liver, colorectal and breast cancers but decreases the risk of prostate cancer.97

The use of metformin, the most commonly prescribed drug for type 2 diabetes, was repeatedly associated with the decreased risk of the occurrence of various types of cancers, especially of pancreas and colon and hepatocellular carcinoma.98

D. OBESITY

Obesity, a rising epidemic, has association with multiple cancers99-101 and has been discussed in detail in an earlier paper in this series. Most of the studies have found an association of increased BMI (marker of obesity) with PCa (RR=1.2-3).84,102,103 A meta-analysis observed 19% increased risk in obese people (BMI>30 kg/m2).101 Obesity related PCa had a population attributable fraction of 26.9% and 19.3% in US and EU respectively.102,104 An earlier age of onset was seen in those who were obese/ overweight during their adolescence (HR=2.09, 1.25- 3.50).102,105 Similarly in the elderly, obesity was found to reduce survival in PCa patients.102 Metabolic syndrome (MetS) is associated with many more consequences than generalized obesity. MetS was found to be associated with PCa in both men (SIR=178,144-266) and women (RR=1.58, p<0.0001).106,107 MetS components were also found to increase the risk [fasting blood glucose (OR=4.24), total cholesterol (OR=1.79), apolipoprotien A (OR=36.06)].108 One European study reported significant risk with several metabolic factors in women (mid-blood pressure, glucose, triglycerides, BMI).109

Physical activity was found to decrease the risk [e.g.: history of sports (HR=0.80, 0.64-0.99), occupational physical activity (RR=0.75, 0.58-0.96)].103,110 Release of cytokines (IL-6, TNF a, CRP) leading to insulin resistance and higher insulin levels result in increased IGF-I: IGFBP-3 (insulin growth factor and binding protein ratio) is probably related to carcinogenesis in obesity.101,106,111

INHERITED PANCREATIC CANCER

Genetic predisposition accounts for 5-10% of all pancreatic cancers.72,87,112,113

Familial pancreatic cancer (FPCA) -is defined as a family with more than one first degree relative (FDR) with history of PCa without any inherited syndromes.112 A 2.3 to 4.5-fold increased risk with 1 FDR, 6.4-fold with 2 FDRs and up-to 32-fold with =3 FDRs with pancreatic cancer has been noted.72,114,115 FPCA which is influenced by race (Ashkenazi Jews), smoking and diabetes, and genetic anticipation (younger age or worse prognosis with successive generations).72 Pancreatic intraepithelial neoplasia with mutations in the K-ras (codon 12) was more frequently (2.75 fold) observed in familial pancreatic cancer when compared to sporadic.72

Hereditary pancreatic cancer- is a genetic syndrome with mutations that increase the risk for PCa.112 Peutz- Jeghers syndrome with STK11/LKB1 gene mutation is associated with up-to 132-fold increased risk for PCa.72,112,116,117 Hereditary Non Polyposis Colon Cancer (HNPCC) is associated with a lifetime risk of 1.3-4% for PCa.112 Majority of the Hereditary Breast Ovarian Cancer (HBOC) is due to mutations in BRCA1 and BRCA2 genes.112 The risk for PCa in BRCA1 carriers is minimally elevated compared to general population (RR=2.8 vs. 1.3%).112 BRCA2 mutation has a 5-7% lifetime risk in carriers and is the most common inherited gene for development of PCa.112 Families with Familial Atypical Multiple Mole Melanoma syndrome (FAMMM) are at increased risk (13-22%) for this cancer.72,112 Individuals with p16/ CDKN2A (FAMMM gene) mutation have a 38-fold higher risk in comparison to general population.72 Studies have shown a mild elevation of risk in FAP with APCA gene mutation (RR=4.5), Cystic fibrosis (CFTR gene) and Ataxia telangiectasia (ATM gene).72,112,116 Associations of PCa with PALB2 (partner and localizer of BRCA2) and palladin (cytoskeleton associated protein) mutations are being observed.72,116,118. (See table X)

OTHER FACTORS

Several studies among different populations across the world have reported an increased risk associated with non-O blood groups for PCa (OR=1.37, 1.02- 1.83).59,119-121 Mechanisms though not clear; relevance of physiological differences in inflammatory mediators (TNFa, cellular adhesion molecules) is being postulated.119 Few studies have found a significant co- relation of this disease with a history of cholelithiasis (HR=3.12, 2.05-4.78) and cholecystectomy (higher prevalence 6.2% vs. 2.9%).108,122 Anti-diabetic medications and NSAID’s are found to have an effect on PCa risk. Metformin (HR=0.73, 0.66—0.80) and thiazolidinediones were associated with reduced risk but insulin (HR=4.63, 2.64—8.10) and sulphonylureas (HR=4.95, 2.74—8.96) aggravate the risk.123-125 DPPV IV inhibitors (sitagliptin) have a theoretical risk for carcinogenesis126 but a recent meta-analysis on this issue reported conflicting results.127

Although not conclusive enough, there is evidence to suggest that high-dose aspirin reduces the risk for PCa (OR=0.78, 0.64-0.95).128-130 Similarly there is lack of satisfactory evidence for other NSAIDs.130 A combination of aspirin, curcumin and sulphoraphane has been found to be beneficial against PCa in animal studies.131 One study from Netherlands observed an inverse association of PCa with hypertension.132 Metformin offers a potential novel approach for pancreatic ductal adenocarcinoma prevention and therapy.133

ALLERGIES

A recent meta-analysis reported 30% drop in pancreatic cancer risk among those with history of allergies.134 Statistically significant risk reduction was observed with hay fever (OR=0.74, 0.56-0.96) and allergy to animals (OR=0.62, 0.41-0.94).134,135 Other allergies, such as those to foods and medications, have been less well studied and associations with risk are unclear.136 Heightened immune surveillance is suggested as the plausible explanation.

GLOBAL EPIDEMIOLOGY

The annual incidence and mortality of PCa is the same (ASR incidence =7.2 vs. 2.8 and ASR mortality=6.8 vs. 2.7).137 Analysis of global data based on human development suggest a higher incidence in areas with high human development (ASR incidence= 4.6) as opposed to areas with less development (ASR incidence= 1.2) (Fig-1).137 Although reasons are not fully elucidated, it is linked to human lifestyle and diet. Immigrant studies, which found increased risk among Indians who migrated to Australia and UK, support this observation.138,139 The highest incidence for women is reported in North American and northern Europe. A high incidence of PCa in Ashkenazi Jews140 and a lower incidence among the Utah Mormons141 has been noted. A brief summary of the incidence of PCa in different parts of the world is given below.

A. AMERICAS

North America, with isolated exceptions, has the highest incidence and mortality for PCa in the world (Incidence ASR= 7.4 and mortality= 6.9) (Fig-1).137 Even though rates in South America are lower, French Guyana and Uruguay are ahead of US and Canada.137 Lowest estimates are seen in Guatemala, Haiti, Panama and the Bahamas (Central America).137

Although PCa ranks only 13th among cancers, it is the 4th major cause of cancer-related death137 with a 5-year survival rate of 6%.142 Blacks have a higher incidence (33% more) and mortality (32% more) than Whites.143 Asia Pacific Islanders have the lowest rates and better survival.142 The rates for the indigenous groups fall between the Blacks and Asia- Pacific Islanders.142 This racial disparity could not be attributed to any of the known risk factors (smoking, BMI, family history, diabetes and cholecystectomy).143 Latitudinal variation in the incidence and mortality of this cancer was attributed to solar UV-B exposure.144

B. EUROPE

Western Europe, Central and Eastern Europe have higher incidence (ASR=6.6-7.3) and mortality (ASR=6.6-6.8) for PCa compared to Northern and Southern Europe (Fig-1).137 Highest incidence is seen in Czech Republic, Slovakia, Hungary, Slovenia and Finland with Czech Republic having the highest incidence rate in the world (ASR=9.7).137 Sweden, Albania, Cyprus and Bosnia have the lowest rates. Hungary has the highest mortality rate for PCa in Europe (ASR=8.8).137

The overall cancer mortality with the exception of PCa has decreased in this region since 1980.145,146 A higher incidence was reported among people living in most deprived areas, partly linked to high prevalence of smoking.146 In England, total number of cases was higher in Whites compared to Non-Whites (South-Asian, Blacks, Chinese 17% of population). However, age standardized incidence was higher in Blacks (ASR=5.7 in Blacks vs. 4.9 in Whites).139 The risk was lower in South-Asians compared to Whites while no significant risk was demonstrated in Blacks and Chinese.139 Immigrant population in England (non- White) had higher incidence of this cancer compared to their counterparts living in homeland.139 A 138% increase in incidence of PCa was reported in the Inuit population (SIR=2.38, 1.97-2.86; p<0.0001) plausibly due to the high prevalence of diabetes and smoking.147 Even though a Nordic country, Finland has a high incidence of PCa (ASR=8.7).137 This is the 5th most fatal cancer in the country (ASR=7.8).137 The 5-year survival has not improved much over the past 50 years (from 3 to 5%).148 Although coffee consumption is very high, it is not associated with an increased risk for PCa (HR=0.82, 0.38-1.76).149

C. ASIA

Eastern Asia (People’s Republic of China, Japan, North Korea, South Korea, Mongolia and Taiwan) followed by Western Asia (Armenia, Azerbaijan, Middle East, Cyprus, Sinai Peninsula of Egypt, Georgia, Turkey) has the highest incidence of PCa in Asia (ASR=4.5 and 3.9 respectively) (Fig-1).137 Similar is the trend in mortality (ASR=4.3 and 3.8 respectively). The Central and South- East Asia have lower rates (Fig-1).137 Highest mortality rate for PCa is in Armenia worldwide (ASR=8.9). Also Armenia has the highest incidence and mortality in this continent followed by Japan, Israel, Kazakhstan and Korea.137 Lowest rates are seen in India, Nepal, Bangladesh, Pakistan and Sri Lanka.137

1. Japan

PCa has a high incidence in Japan (ASR=8.5) and is the 5th common cause of cancer-related mortality in both men and women (ASR=9.5 and 6.1 respectively).137 Although the 5-year survival is around 5%, resected cases have better prognosis (5-year survival=18.8%).150 Positive family history and presence of diabetes were reported as major risk factors apart from smoking.151 Northern Japan has a higher mortality from this cancer as opposed to south which was linked to variations in exposure to solar irradiation and temperature.152 A high non-linear relationship of PCa death with low- dose external irradiation (<20mSv) was reported in Japanese-A bomb survivors.153

2. Korea

With a high incidence (ASR=6.7), PCa in Korea is the 5th most fatal cancer (ASR=6.2).137 Analysis of trends in incidence from 1999-2010 showed that PCa is increasing in Korea (APCA in both sexes=1.4) with a greater increase in women (APCA=2.2 vs. 0.6 in men).154 Although hepatitis B is endemic in Korea, it was not found to increase the risk for this malignancy (OR 1.03, 0.69-1.53; p=0.91) as opposed to hepatitis C (OR=2.30, 1.30-4.08; p< 0.01) and non-O blood group (OR=1.29, 1.05-1.58; p=0.01).59

3. China

The incidence of PCa is not very high in China (ASR= 3.6) but it is the 8th most common cause of cancer related death.137 Cigarette smoking (44%), pancreatitis (16%) and family history of PCa (8%) were the major etiologies in young patients with PCa.155 Incidence of diabetes among PCa patients was much higher (34.6% vs. 8.8%) and 74.56% (in the cancer group) had onset of DM within 2 years of diagnosis of cancer.156 Energy dense foods increased the risk (OR: 1.72; 95% CI: 1.25, 2.35; P = 0.001) in Chinese.28 Regular green tea drinking was found to be protective in Chinese women (OR 0.68, 95% CI 0.48—0.96).157 For reasons not clear, in one study acute pancreatitis patients had a high risk of developing PCa within 5 years of index pain (HR=9.10; 3.81-21.76).66

4. India

India has a low incidence (ASR=1.2) and mortality (ASR=1.1) for PCa.137 Recent studies note an increasing incidence of PCa in both men and women.158 A greater risk was observed in educated males with about 3-fold risk in those with >12 years of education.159 Tropical calcific pancreatitis, which has a very high risk for this cancer (RR=100, 37-218), is highly prevalent in some parts of India.73,75

5. Israel

PCa in Israel ranks 3rd in men (ASR=8.6) and 4th in women (ASR=6.2) in malignancy related deaths.137 Jews had a higher incidence than Arabs (ASR males 7.45 vs. 5.61) with the highest incidence (ASR males 8.11 vs. 7.45) in immigrant Jews (European-born).160 Nevertheless, a decreasing trend is seen in the Jewish population160 Mutation in BRCA1/2 was the major cause of this cancer in Ashkenazi Jews.161

D. AFRICA

The southern part of the continent, which includes South Africa, Mauritius and Zimbabwe, has the highest estimates for PCa in Africa (ASR incidence=4.3, ASR mortality=4.2) (Fig-1).137 Libya and Egypt, though in the northern part has incidence and mortality rates similar to Southern Africa.137 There are few studies, which attribute pollution of the Nile for the increased incidence in Egypt.43,162 Serum cadmium is suggested as the etiologic agent for the occurrence of early- onset PCa in the East Nile delta region.43 A Moroccan study reported higher incidence (17%) of pancreatic adenocarcinoma in adults <45 years (3% in the US) which did not correlate with smoking, alcoholic pancreatitis or family history.163 Lowest incidence (ASR<2) and mortality (ASR<1.9) for this malignancy is seen in Malawi, Guinea and Tanzania.137

E. OCEANA

Australia has the highest incidence (ASR=6.6) of PCa in this region followed by New Caledonia (ASR=6.5) and New Zealand (ASR=5.9) while island countries such as Samoa and Vanuatu have the least (Fig-1).137

1. Australia

PCa is the 5th common cause of cancer-related death in Australia.137 A rise in incidence (ASR 7.67 to 8.24) and mortality (ASR 7.02 to 7.58) of PCa was observed in women (from 1977-2006) while these estimates dropped in men164 attributed to variation in smoking habits.164 An interesting observation is that mortality from PCa was 9% less in Brisbane (Queensland) when compared to Melbourne (Victoria), which was linked to variations in UV exposure between the two capital cities.165

2. New Zealand

PCa is the 5th most fatal cancer in this country.137 A higher incidence and worse prognosis was reported among the Maori tribe which was attributed to smoking.166 Blakely et al. analyzed cancer incidence in New Zealand by dividing the population into 4 ethnic groups (Maori, Pacific, Asian and European/others) and found a 1.5-times higher rates for PCa in the Maori and Pacific groups compared to the European group.167

CONCLUSION

PCa continues to be a major clinical challenge because of a trend in increase in incidence with no major improvements in survival. Recent studies have reported miRNA-21, c-Myc, L-type amino acid 1 transporter (LAT1), K-ras codon 12 mutation, p38? Mitogen-activated Protein Kinase and SMAD4 as biomarkers in predicting prognosis and survival in these patients.168-174 Circulating Tumor Cell (CTC) detection in peripheral blood with a diagnostic accuracy of 70% (EUS-FNA=85%) is a promising noninvasive early diagnostic procedure.175 A study from MD Anderson, Texas reported an association between NAFLD and pancreatic cancer.175 If this observation is confirmed in further studies, it is concerning since MetS and obesity are rapidly increasing. Furthermore they observed simultaneous pancreatitis and liver cirrhosis in obese pancreatic cancer patients providing additional evidence for the role of obesity.175

Endocrine (islet cell tumors) and rare non-endocrine tumors (acinar cell carcinoma, adenosquamous carcinoma, colloid carcinoma, giant cell tumor, hepatoid carcinoma, intraductal papillary-mucinous neoplasm, mucinous cystic neoplasm, pancreatoblastoma) of the pancreas are not discussed in this article.

Regardless of the advances in medical science, PCa remains a challenge. More desirable survival outcomes rely on novel research that focuses on finer diagnostic and therapeutic approach, yet to materialize.

Download Tables, Images & References

A CASE REPORT

Bezoar in a Periampullary Duodenal Diverticulum Causing Pancreatobiliary Obstruction

Read Article

In this case, we report a bezoar filling a duodenal diverticulum causing obstruction of the pancreatobiliary tree in a patient presenting with severe epigastric pain. Duodenal diverticula are typically asymptomatic outpouchings that form due to chronic duodenal ulcers or laxity of the bowel wall around the ampulla. When symptomatic, patients usually present with abdominal pain caused by duodenitis. People with gastroparesis, prior gastric surgery or a large diverticulum are at increased risk of forming a bezoar that can lead to gastric, small bowel, biliary or pancreatic duct obstruction. In a patient presenting with epigastric pain, laboratory results and cross sectional imaging are the quickest and easiest way to help make the diagnosis of an obstructing bezoar. Treatment options include conservative management, endoscopy or surgical removal of the obstructing mass.

Javier Rivera, MD, PGY-3, Diagnostic Radiology Paul Klepchick, MD, Diagnostic Radiology, Division of Abdominal Imaging Deepti Dhavaleshwar, MD, PGY-5, Department of Gastroenterology Allegheny General Hospital, Pittsburgh, PA

A 56 year old female with a 40 pack-year smoking history and a past medical and surgical history of depression, appendectomy and cholecystectomy presented with abdominal pain and 3-4 episodes of nonbloody and nonbilious emesis. She was hospitalized 3 times over the course of 3 weeks for pancreatitis at an outside hospital. Outside images were not available, however, a magnetic resonance cholangiopancreatography (MRCP) performed one month prior to the current presentation reported extrahepatic and intrahepatic ductal dilation, without choledocholithiasis. A planned endoscopic retrograde cholangiopancreatography (ERCP) was cancelled at the patient’s discretion. A chest computed tomography angiogram (CTA) was performed on a prior hospitalization due to symptoms related to congestive heart failure or possible pulmonary embolus. The CTA reported no pulmonary embolus; however, it did report small bilateral pleural effusions, bibasilar atelectasis and moderate emphysema. On the current admission, her symptoms included sharp and constant epigastric pain radiating to her back, rated 10/10 that was relieved by narcotics. She denied fevers and abnormal bowel movements. Vital signs at presentation were: temperature 37.1, heart rate of 75 bpm, blood pressure of 117/67 mmHg, respiratory rate 17, oxygen saturation 97% on room air. Physical exam revealed a neurologically intact, slightly jaundiced female in mild discomfort. The abdomen was non-distended and bowel sounds were noted in all four quadrants. The abdomen was tender in the epigastric region to light and deep palpation. No rebound tenderness was noted, nor ascites, palpable mass or costovertebral angle (CVA) tenderness. Laboratory findings include alanine aminotransferase level of 472 U/L, alkaline phosphatase level of 832 U/L, aspartate aminotransferase level of 275 U/L, lipase level of 1461 U/L, total bilirubin level of 2.9, WBC level of 12.2 k/mcL.

A CT scan of the abdomen and pelvis was performed and the patient was referred to the gastroenterology service. The patient was given one dose of 400mg ciprofloxacin and 500mg of flagyl in the emergency department due to the mildly elevated white blood cell count and was continued on the same antibiotic regimen when admitted. She was placed on narcotics and anti- emetics and was scheduled for an ERCP.

Radiographic and ERCP findings

The CT scan showed an obstructing bezoar within a 4.0 x 3.7 cm periampullary diverticulum causing severe intrahepatic and extrahepatic biliary dilatation, pancreatic duct dilatation and mild fat stranding surrounding the pancreatic head (Figures 1-2). Differential diagnosis included duodenal diverticulitis/ abscess or obstructing malignancy.

The ERCP showed two diverticula at the major papilla. The papilla was at the rim of a single diverticulum, which was ulcerated and impacted with solid food (Figure 3). The majority of the food was removed using a tripod and multiple flushes through a cannula. After clearance of the bezoar, a gush of bile followed the relief of the obstruction. Sphincterotomy was not performed due to the edema caused by the bezoar. The patient improved after treatment.

A follow up ERCP was done at 5 weeks. A single peri-ampullary diverticulum with a small opening was found at the major papilla, without evidence of bezoar impaction. The biliary duct was cannulated with a 44 sphincterotome. Severe, diffuse dilation of the biliary tree was present, with the CBD measuring 15mm. A balloon sweep of the CBD was performed using a 15- 18mm balloon, without evidence of stones, sludge or debris. An endoscopic ultrasound (EUS) was also performed at this time, which revealed a normal pancreas and pancreatic duct, with no evidence of divisum. Follow up laboratory values also improved. Alanine aminotransferase level of 14 U/L, alkaline phosphatase level of 115 U/L, aspartate aminotransferase level of 18 U/L, lipase level of 85 U/L, total bilirubin level of 0.3, WBC level of 10 k/mcL.

DIAGNOSIS

Periampullary duodenal diverticulum with bezoar causing pancreatobiliary obstruction.

DISCUSSION

The most common causes of biliary obstruction is a common bile duct stone. Other causes include pancreatic carcinoma, cholangiocarcinoma, stricture of the CBD, infection or recent instrumentation.

A search of the current literature3-6 revealed that bezoars are uncommon sources of bowel obstruction, and are more rare causes of pancreatobiliary obstruction and acute pancreatitis. A bezoar is an indigestible mass that commonly develops in the stomach; however, they can also form in the small bowel. Categories of bezoars include phytobezoar, made predominantly of plant matter; lactobezoar, made of undigestible milk; pharmacobezoars, made of medications; and trichobezoars, found in people with tricotillomania who present with an obstructing conglomeration of hair in Repunzel syndrome. People with prior gastric surgery and gastroparesis are at increased risk of bezoar formation.5 Surgeries that bypass the pylorus may cause large food boluses to enter the small bowel, which could potentially cause small bowel obstruction. Stasis of foodstuff in the stomach, or in this case, a diverticulum, may also predispose people to forming bezoars and obstruction. Treatment of bezoars in the stomach and duodenum include ERCP or surgery, if the bezoar cannot be removed with a net or basket. If the bezoar is causing a small bowel obstruction, patients are managed surgically with enterotomy.

A recent case report described bezoar-induced pancreatitis secondary to a bezoar in a periampullary diverticulum, however there was no biliary obstruction.7

A periampullary diverticulum contains or is adjacent to ampulla of Vater, while a juxtapapillary diverticulum originates within 2cm from the papilla. Diverticula may form due to wall weakness where the common bile duct (CBD) and pancreatic duct (PD) attach to the duodenum or with increasing age. Duodenal diverticula are typically asymptomatic, and are found incidentally in 6% of Upper GI studies and up to 23% at autopsy.1 They can, however, cause sphincter of oddi incompetence, abdominal pain, jaundice, bleeding, biliary stones, diverticulitis, perforation and pancreatic or CBD obstruction. In our case, the diverticulum filled with a food bolus. Cross sectional imaging with oral and intravenous contrast has helped in the detection of air fluid levels within the diverticulum, identifying the diverticular neck and sequela of possible obstruction of the small bowel or pancretobiliary system.7


Treatments of periampullary diverticula include
conservative management, endoscopy or surgical
removal.

Download Tables, Images & References

NUTRITION ISSUES IN GASTROENTEROLOGY, SERIES #137

Clinical Observations Correcting Hypernatremia: Enteral or Intravenous Hydration?

Read Article

Carol Rees Parrish MS, RD, Nutrition Support Specialist, Digestive Health Center of Excellence, Mitchell H. Rosner, MD, Henry B. Mulholland Professor of Medicine Chairman, Department of Medicine, Division of Nephrology, University of Virginia Health System, Charlottesville, VA

BACKGROUND

Hypernatremia is defined as a serum sodium value > 145 mEq/L. Since the serum sodium is determined by the ratio of the amount of sodium in the serum to the amount of plasma water, hypernatremia can develop from either an excess of sodium (such as due to the administration of hypertonic fluids), a loss of hypotonic fluids (free water) or a combination of both. Most commonly, it is the loss of hypotonic fluids and the failure to replace these water losses that result in hypernatremia.

In most circumstances, thirst is a powerful defense mechanism against a rise in the serum sodium level. The body defends its serum osmolality closely so that as the serum sodium rises (and with it serum osmolality), thirst ensues along with rises in arginine vasopressin (AVP) secreted by the posterior pituitary. AVP leads to urinary concentration and conservation of renal water excretion, but ultimately it is thirst and ingestion of water that allows the serum sodium to normalize. Thus, most patients who develop hypernatremia have the common feature that water intake is restricted in some form. For instance, patients in the intensive care unit (ICU) who are intubated and sedated cannot control their water intake and the same is true for patients with impaired mental status or limited mobility. Thus, these patients are at high risk for hypernatremia (Table 1).

Epidemiology of Hypernatremia

Given the powerful ability of thirst to defend against hypernatremia, it is not surprising that the incidence of this electrolyte disorder in patients presenting to the emergency department (ED) is uncommon (0.2%).1 Most of these outpatients usually have either chronic or acute impairment in their mental status (such as dementia). In critically ill patients, the incidence of hypernatremia is 10-fold higher (2-6%).2,3 Importantly, a large percentage of patients develop hypernatremia during the course of their hospital stays (especially in the ICU – up to 10%).4,5 The reasons for this are multi- fold and include:

  • Use of hypertonic fluids
  • Ongoing loss of body fluid loss (gastric decompression, stool, ostomy, fistulas, biliary drains, etc., inattention to water intake and needs)
  • Use of diuretics, lactulose
  • Poorly controlled hyperglycemia.6

Consequences and Complications Associated with Hypernatremia

The clinical symptoms associated with hypernatremia are alterations in central nervous system functioning, including a spectrum ranging from mild confusion to stupor and coma. These symptoms likely result from changes in cellular volume as water moves from the intracellular compartment to the more hypertonic extracellular compartment resulting in cell shrinkage.7

Mortality rates in patients with hypernatremia, especially those in the ICU, are very high (ranging from 15 to 50%), depending upon the severity of the hypernatremia.8,9 While hypernatremia has an independent effect on increased mortality, the underlying disease processes driving the development of hypernatremia is more likely to blame with the higher mortality rates.10

Etiology of Hypernatremia

Broadly speaking, the causes of hypernatremia can be divided into three categories:

  • water with solute loss (with water losses in excess of solute losses)
  • pure water losses
  • solute (sodium) gain.11

In those cases of water with solute loss and pure water losses, most patients will have impaired mental status and decreased thirst sensation, or the inability to obtain free water (see Table 2).

Correction of Hypernatremia

As many patients with hypernatremia will be volume depleted as well as dehydrated, assessing the need for rapid resuscitation is critical, and if needed, intravenous isotonic solutions should be administered until the patient is hemodynamically stable. Before correction of hypernatremia, it is vital to determine if the rise in serum sodium is acute (< 48 hours) or more chronic (> 48 hours). This is because with chronic hypernatremia, brain adaptations in cellular volume have occurred such that rapid correction in these circumstances can result in cerebral edema, increased intracranial pressure and brain stem herniation with death.12 In chronic states of hypernatremia the serum sodium should not be lowered by more than 8 to 10 mmol/L/24 hours. If it is unclear as to the duration of hypernatremia, it is best to assume that the condition is chronic and use a slower rate of correction.

Several formulas are available to determine the rates of infusion of hypotonic solutions and any of these can be utilized.7,13,14 However, it is critical that the clinician measure serum sodium levels frequently the duration of correction (every 4-6 hours), so that over- rapid correction is avoided and infusion/replacement rates for water can be adjusted. Water replacement can be achieved in several manners: (1) intravenous hypotonic fluids which may range from 0.45% saline to 5% dextrose in water or (2) enteral water administration.

In many cases, it is important to recognize that the free water deficit may be great and along with continuing water losses (insensible losses), the replacement rates can be substantially greater than 1-2 liter per day. When replacement rates are greater than 1 liter in a 24 hour period, administering water replacement only via the gastrointestinal route may prove challenging as the following cases highlight.

CASE 1

46 year-old female who is post-repair of a congenital heart defect is now requiring mechanical ventilation. A percutaneous gastrostomy (PEG) tube is in place and the patient is tolerating enteral nutrition (EN) with a daily volume of 1400 ml of enteral formula. Over the past few days, she has become hypernatremic and the care team increased her water flushes to 300mL every 4 hours (1800mL per 24 hours) to address her estimated water deficit. Thus, in total, she was receiving 3200 ml of fluid enterally per day. In this case, she became quite uncomfortable and given this concern as well as concerns for gastric distention and risk for aspiration, it was recommended she be given IV replacement of her free water deficit and minimize enteral delivery to only tube feeding and necessary medications until the distension resolved. Two days after the switch to IV free water replacement, her abdominal distention had resolved.

CASE 2

A 72 year-old male admitted to the intensive care unit (ICU) with sepsis developed severe diarrhea due to Clostridium difficile colitis. He is receiving enteral nutrition, but due to ongoing diarrhea and fluid losses he became hypernatremic. The patient was prescribed 400mL of water every 3 hours to replace the water deficit. With this increased fluid delivered enterally, he vomited leading to a change to IV water replacement.

Practical “Water Replacement”

The GI tract is normally capable of handling a large amount of fluids and solute/nutrient administration. However, in the setting of critical illness or serious GI problems (gastroparesis, etc.), this capacity for water and nutrient adsorption may be compromised and lead to increased abdominal distention, discomfort and the risk for nausea/vomiting and aspiration events. Water given the enteral route vs. the intravenous route is equally as effective; however there may be practical limitations to the use of the GI tract especially when replacement volumes are greater than 1000mL. On the other hand, IV water replacement is safe, reliable, predictable, consistent, and avoids these complications.

Oral versus IV Hydration

If the clinician is contemplating oral hydration, it is important to assess the ability of the GI tract to accommodate and absorb the additional fluid load. This is especially true if the patient is already receiving EN and the additional volume with hydration may prove intolerable. Certainly, if a patient is having trouble tolerating EN at the outset, then IV hydration should be the route of choice. Other key questions that need to be asked before utilizing enteral hydration include:

  1. Will enteral solutions be held for periods of time (such as for procedures, lost access)?
  2. Does the patient have impaired GI motility?
  3. Is the patient constipated?
  4. Is the patient at risk for aspiration that could be worsened by increased GI distention?

Finally, it should be noted that when large and frequent water flushes are ordered (such as 300mL every 4 hours), each time the flush is scheduled to run in, enteral feedings are stopped or automatically shut off (if the patient is on a dual pump), while the water infuses. The larger the flush, the more time it takes to run in. If the patient has frequent and large flushes, this can translate into significant lost feeding time and the patient?s nutritional status is compromised. Furthermore, if the above dual enteral pump is being used and the patient is on a nocturnal, or other cycled delivery mode, when the pump is turned off, no flushes will be infused.

Suggested Guidelines

  1. For serum sodium < 150 mmol/L, it is reasonable to try enteral water replacement up to 1 liter in divided doses (for example, 250mL every 6 hours, or 165mL every 4 hours).
  2. For serum sodium > 150 mmol/L, IV hydration should be given carefully, and in a controlled and reliable fashion, using dextrose 5% in water or another hypotonic fluid as appropriate for the individual patient.

Download Tables, Images & References

NUTRITION ISSUES IN GASTROENTEROLOGY, SERIES #136

Blenderized Tube Feeding: Suggested Guidelines to Clinicians

Read Article

Blenderized tube feeding has been gaining momentum among patients despite the availability of commercially prepared and nutritionally complete enteral formulas. This article will review perspectives and provide clinicians with basic guidelines for planning and implementing blenderized enteral feedings when our patients request this feeding option.

INTRODUCTION

Before the availability of commercial enteral formulas, “blenderized” foods were prepared in hospital kitchens to create liquid mixtures given by bolus syringe method through large-bore nasogastric and gastric tubes. As technology continued to advance in the 1970s, commercial formulas of defined composition were introduced for tube feedings.1 Blenderized tube feeding (BTF) became a less desirable option when commercial formulas became more affordable, easy to use, and widely available. The risk of microbial contamination and labor intensity were the primary reasons BTF were abandoned by most healthcare facilities. Commercial enteral formulas are packaged aseptically and are approved to hang for 48 hours as a closed system reducing administration error and time involved with tube feeding. Unfortunately, the only tubing available in the U. S. for use with these products cannot hang for more than 24 hours (so this touted 48 hour benefit is not really a benefit). Studies have demonstrated that the microbial quality of hospital- prepared BTF was not within the published guidelines for safety.2-3 However, in many parts of the world, especially in developing countries, the use of hospital- prepared formula is a routine practice due to economic or cultural reasons.2-4

Compared with commercial formulas, preparation of BTF can be labor intensive and the cost (is rarely, if ever) covered in medical plans. In addition, there are food safety concerns and uncertainty about the nutritional value of non-standardized recipes. For these reasons, clinicians are hesitant to recommend this option to an already stressed and exhausted patient and /or caregiver. Several studies have shown that hospital- prepared formulas provided unpredictable levels of macro and micronutrients, delivered less than the desired amounts of nutrients, and increased the risk of tube occlusion due to viscosity, making it unsuitable for reliable infusion through feeding tubes.2,4

However, there remains a subset of the population who are interested in this feeding option. BTF has been gaining popularity among home enteral nutrition (HEN) patients, particularly the pediatric population. Many patients and families choose BTF because of perceived health benefits, intolerance to commercial feeding formulas, food allergies, improved bowel function, psychosocial reasons, or personal preference (desire for “real” food, organic, vegetarian, etc.).5-6

Use of Blenderized Tube Feeding

BTF is defined as the use of blended foods and liquids given directly via the feeding tube. Historically, these types of enteral formulas have been called “blended diet”, “blenderized feeding,” “blended formula,” or “homemade blended formula.” The practice of incorporating BTF into a feeding regimen could be a combination of a commercial formula and commercial pureed baby food, or three meals a day of homemade blenderized foods supplemented by nocturnal feeding of a commercial formula, or complete feedings using homemade BTF made from recipes, and many variations in between. The introduction of ready to use BTF in the marketplace has also provided HEN patients with an option over commercial enteral formulas. Regardless of how a patient uses BTF, it is essential to identify a commercial enteral formula for emergency situations, or when traveling when refrigeration is not available.7

There is little published research available to support the efficacy of BTF that translate into any type of beneficial outcomes of this feeding technique. There are numerous anecdotal reports from patients, caregivers, and medical professionals of positive experiences that have been shared through informal patient questionnaires, feeding support groups on the internet, social media, professional discussions and clinical experiences.5-9

One feeding clinic reported both medical and emotional benefits from the use of BTF.7 It allowed for some normalization of the feeding process for gastrostomy tube-fed patients, greater volume tolerance, and improvement in reflux and constipation, and it facilitated the transition from tube to oral feeding. The use of blenderized foods allowed for inclusion of a tube-fed patient in family mealtimes and a sense of “normalization” of gastrostomy tube feedings. It also promoted the view of the G tube as another mouth, thereby priming the gastrointestinal system for the complexities of food.

The interdisciplinary feeding team at the Cincinnati Children’s Hospital Medical Center (CCHMC) conducted a feeding trial using a pureed diet given directly into the feeding tube, referred to as the pureed by gastrostomy tube (PBGT) diet. It was designed for children with gagging and retching after fundoplication surgery.8 This was the first clinical trial using BTF to manage the complications associated with enteral feeding. Fifty-two percent of parents reported a decrease in gagging and retching after their child started the PBGT diet and 57% of children were reported to have an increased oral intake.8

When to Consider Blenderized Tube Feeding

Table 1 outlines the prerequisites before seriously considering the use of BTF. Ideally, a clinician must first determine if a patient is a good candidate for BTF. However, a tube-fed patient may already have transitioned to either partial or full homemade BTF regimen before a referral is made. Working with a dietitian is essential to ensure that the homemade diet is adequate and whether the current homemade BTF recipe plan needs modification. The best candidate is a patient and/or caregiver who made the decision to “try” this feeding option, and is willing to commit their time and effort for instruction and preparation of BTF. A patient who is having tolerance issues or allergy to a commercial enteral formula may also be a candidate after discussion with the medical team and dietitian.

When Not to Consider Blenderized Tube Feeding

BTF is not an option for all tube-fed patients. Patients with complicated medical and gastrointestinal issues and those who require frequent hospitalization may not tolerate and sustain a BTF regimen. Often, these patients require specialized enteral formulas. A patient who requires continuous feeding is not a good candidate since a homemade blended formula is not recommended for feedings that will last for more than 2 hours due to concerns over food safety and bacterial contamination.5

Homemade BTF is generally thicker and can potentially clog the feeding pump making it difficult to flow through the feeding set. Some patients have successfully infused BTF through a feeding pump as long as the mixture is thinned with additional fluid, blended, and strained sufficiently. Bolus syringe method works best and provides the pressure needed to move a homemade blended formula down a feeding tube. It is recommended for gastrostomy tubes 14 French size or larger to prevent clogging and for ease of administering the diet.

Tools for Success

  • Heavy duty blende
    • Blendtec® HP3 blender: www.blendtec.com (800) 253-6383
    • Vitamix®: Inquire about Vitamix® Medical Needs Discount Program which is available to all eligible candidates at (800) 848-2469 or email: household@vitamix.com reference code 07-0036-0011
  • Strainer or fine sieve if using a regular kitchen blender or stick blender (see Figure 1)
  • Airtight storage containers, ice cube trays for freezing individual portions
  • Adequate refrigeration/ freezer space
  • 60 mL syringe with plunger
  • Bolus extension set for low-profile gastrostomy tube
  • Feeding pump (if using)
    • Make sure BTF is thin enough to flow easily through the pump
    • Discard BTF after 2 hours maximum if kept at room temperature for that long
  • Insulated bag or ice chest with ice packs when traveling
  • Patient education on food safety guidelines (www.fsis.usda.gov, www.foodsafety.gov, www.homefoodsafety.org)

Getting Started with a Homemade Blenderized Tube Feeding

  • Evaluate the patient’s medical history, success (or not) with current feeding regimen (feeding tube, oral intake), food tolerances, lifestyle/ethnic/religious preferences, ability to obtain individual ingredients and tools needed for preparing a homemade BTF.
  • Discuss with medical team and determine if patient is ready to start the transition. Most patients transition slowly and use a commercial formula for part of nutrient requirements.
  • Determine goals for calories, protein, fluids, and vitamin, mineral, and electrolyte supplementation.
  • Develop a meal plan and starter recipe. The following tools can be used as starting point to create a meal plan based on caloric goals:
    • USDA Choose My Plate http://www.choosemyplate.gov/ supertracker-tools/supertracker.html The supertracker feature provides individualized worksheets where a profile can be created to calculate and track a menu plan. A sample 1000 calorie meal plan will include:
      • Grains: 3 servings
      • Fruits: 1 serving
      • Vegetables: 1 serving
      • Protein (meat, beans, or nuts): 2 servings
      • Milk or milk substitute: 2 servings
      • Fats: 3 servings
    • Homemade blended formula worksheets by Dunn Klein M, Morris SE. Homemade blended formula handbook. Mealtime Notions, LCC, Tucson, AZ, 2007; 117-128; www.mealtimenotions.com The worksheets were adapted from the USDA My Pyramid.
    • Sample Blenderized Tube Feeding Recipes at www.ginutrition.virginia.edu under Resources for Nutrition Support Clinicians
    • Nutrition and recipe analysis applications. Computerized nutrition programs such as Food Processor®, Nutritionist ProTM allows the clinician to add and modify foods while monitoring the total caloric and protein levels along with vitamin and mineral profile. The clinician can save and retrieve the data for future reference.
    • USDA National Nutrient Database for Standard Reference: http://www.nal.usda. gov/fnic/foodcomp/search
    • Exchange Method
    • Food company websites/food labels
  • Determine macronutrient food sources. The following examples work well in blenderized tube feedings:
  • Grains: cooked cereals, boiled white or brown rice, cooked quinoa, oats, regular or whole grain bread
  • Fruits: avocado, applesauce, peach, pear, banana, papaya, blueberries, 100% fruit juice (pulp free). Commercial pureed baby food (stage 2) can be used for variety, consistent nutritional value, and it also avoids the potential complication of clogging the feeding tube. It is convenient especially when traveling, and avoids the need to purchase an expensive heavy-duty blender. The patient can eventually advance and transition to blenderized table foods.
  • Vegetables: white potato, sweet potato, carrots, squash, well-cooked broccoli. Can use commercial pureed baby food (stage 2).
  • Protein: chicken, beef, legumes, soft tofu, smooth 100% peanut butter, cooked eggs, canned tuna or other fish without bones. A commercial enteral formula can be used as the base for a blended diet instead of meat, milk, or yogurt as the sole protein source (see Table 2). Milk and yogurt tend to blend more easily than cheese.
  • Milk or milk substitute: cow’s milk, soy milk, almond milk, rice milk, yogurt, non- fat milk powder.
  • Fats: canola, olive, flaxseed, hemp and corn oils.
  • Example of a 1000 calorie blend:
    • Grains: 1 cup cooked oatmeal and 1/2 cup cooked brown rice
    • Fruits: 1/2 cup peaches (canned or fresh) and 1/2 cup unsweetened apple juice
    • Vegetables: 1/2 cup cooked carrots and 1/2 cup cooked sweet potato, butternut squash or pumpkin
    • Protein: 2 ounces cooked chicken
    • Dairy: 1 cup whole milk and 1 cup yogurt (plain)
    • Fats: 3 tsp olive oil
  • Review and modify the recipe using a recipe analysis program to adjust the ratios of macronutrients to the desired composition.
    • A modular product (glucose, MCT oil, protein, fiber such as Benefiber®) can be added to enhance the nutrient profile of blenderized enteral feedings.
  • Determine fluid goals. The percentage of free water is calculated from the blended recipe by multiplying the total volume of fluid- containing ingredients (commercial formula, meats, fruits, vegetables, milk, and yogurt) in ounces by 0.75. This is based on the assumption that most infant foods contain roughly 75% free water (9). Additional free water is determined from the difference between estimated daily fluid requirement and amount of free water of the blended recipe.
  • Determine need for vitamin, mineral, and electrolyte supplementation. Perform a recipe analysis. Request labs as appropriate for any nutrients of concern. A multivitamin may be crushed and added to one of the bolus feeds. A liquid multivitamin is an option, but is not always complete – make sure patient gets the right one. Oral rehydration solution (ORS) can be given instead of free water flushes between bolus feeds.
  • Patient and/or caregiver education
    • meal plan, ingredients, starter recipe
    • equipment: blender, strainer, 60 mL syringe, storage containers
    • preparation details
    • proper sanitation method
    • proper storage and refrigeration
    • administration method and feeding schedule (bolus syringe, pump)
    • water or oral rehydration solution flushes
    • vitamin/mineral/electrolyte supplementation if needed
    • travel/emergency plan
  • Patient monitoring and follow-up. Schedule call-back, follow-up visit to monitor weights/ weight changes, modify/adjust BTF recipe.

Commercial Ready to Use Blenderized Tube Feeding Products – See Table 3

  • Liquid HopeTM. Ready to use organic whole foods. The manufacturer recommends switching to Liquid HopeTM slowly, one meal per day until the transition is complete. The unopened formula is shelf stable for up to 2 years. It has a 3-hour room temperature hold time and unused formula can be refrigerated for 48 hours. The formula can be diluted with water to achieve the desired consistency. Patients can order online ($7.99 per 12-oz serving; available in units of 6, 12, or 24). Medicare approved code for this product is B4149 (blenderized natural foods with intact nutrients). http://www.functionalformularies.com
  • Real Food BlendsTM. Pre-made blenderized meals, available in 3-meal varieties (Salmon Oats and Squash, Orange Chicken Barley and Carrots, Quinoa Kale and Hemp) and is shelf stable for 18 months. Unused formula should be refrigerated and used within 24 hours. The quinoa meal is slightly thicker and may require additional fluid to flow easily. Patients can order online ($49.95 for 12-pack meals); approved by CMS HCPCS for code B4149 (blenderized natural foods with intact nutrients).
    Discover 100% Real Food for Tube Feeding with Real Food Blends
  • Compleat©. Formulated with real food ingredients: chicken, peas, carrots, tomatoes, and cranberry juice. Available in 250 mL carton and 1 liter closed system. HCPCS code B4149 (blenderized natural foods with intact nutrients). http://www. nestlehealthscience.us/products/compleat

Use of Blenderized Tube Feeding and the New ENFit Connectors

The new enteral connector (ENFit) system is being introduced in phases and is expected to be completed in 2015. GEDSA (Global Enteral Device Supplier Association), Kimberly-Clark, and A.S.P.E.N. (American Society for Parenteral and Enteral Nutrition) recently conducted experiments to determine the pressure required to dispense a BTF through a 60 mL syringe (catheter tip and ENFit) and to check the gravity flow (mL/min) of BTF through the connectors.10 The BTF was tested just coming out of the refrigerator to simulate clinical conditions. It took about the same pressure to push the BTF through the catheter tip and ENFit syringe, but was actually easier with the ENFit connector. The testing also showed that the gravity flow (mL/min) of BTF for ENFit and catheter tip syringes were essentially equivalent through two gastrostomy tubes (Kimberly-Clark 18fr/20fr PEG with solid bolster and balloon retained G tube feeding systems).

SUMMARY

The process of transitioning to BTF can be an overwhelming task for both the patient and clinician. Homemade blended diets are often described as nutritionally unbalanced, have increased risk for food contamination, and may compromise the enteral access devices. BTF can be used for partial, supplemental, or complete nutrition support. It can be safely used and implemented with the involvement of the medical team and support from the RD to assess, educate, and monitor the patient’s progress toward this feeding option. See Table 4 for additional resources.

Download Tables, Images & References

DISEASES OF THE PANCREAS, SERIES #9

The Physiology of the Pancreas

Read Article

Here we discuss the pancreas, a complex organ that plays a critical role in the digestion process. Knowledge of normal pancreatic secretory functions can help clinicians order appropriate tests, which eventually assist in diagnosing specific pathologies. New imaging techniques such as computed tomography (CT) and magnetic resonance imaging (MRI) have enhanced the understanding of pancreas anatomy and made possible the surgical, endoscopic and percutaneous manipulation of pancreas.

INTRODUCTION

When originally studied, the pancreas was one of the last organs in the abdomen to catch critical attention of anatomists and physiologists.1 New imaging techniques such as computed tomography (CT) and magnetic resonance imaging (MRI) have enhanced the understanding of pancreas anatomy and made possible the surgical, endoscopic and percutaneous manipulation of pancreas.2

Anatomy

The pancreas is a flat, long, and soft gland that is roughly 6 inches in length and weighs between 70 and 110 g. It lies obliquely in the retroperitoneal space of the upper abdomen and is covered by the stomach, transverse colon, and transverse mesocolon. The head of the pancreas lies next to the duodenal sweep at the level of the body of L2.3,4 The neck, body, and tail span leftward, with the tail extending close to the spleen.5

Functional Anatomy

The pancreas consists of ≥80% of acini that are arranged in clusters that form lobules separated by loose connective tissue. A circular shaped acinus and its tubular draining ductule form the functional unit of exocrine pancreas.6 Many acini arranged like a bouquet secrete digestive enzymes into the ductule which drains into interlobular ducts and finally into the main pancreatic ductal system.

Pancreatic acinar cells are specialized exocrine secretory cells that synthesize, store, and secrete the digestive enzyme component of the pancreatic juice. An acinar cell is shaped like a triangle, with the basal membrane pointed outward for neurohormonal receptors and the apical membrane located inward forming the lumen of an acinus. The nucleus and rough endoplasmic reticulum (RER) are located near the basal membrane for protein synthesis.7 Zymogen granules that store digestive enzymes are located near the apical membrane and hence close to the lumen. Tight junctions between acinar cells form a barrier between the lumen and apical membrane to prevent inappropriate passage of enzymes but allow water and ions to go through.8,9 Secretagogues stimulate acinar cells causing the granules to fuse with each other and the apical membrane. Microvilli covering the apical surface of acinar cells facilitate exocytosis of enzymes into the lumen. Gap junctions between adjacent acinar cells allow coordinated chemical and electrical communication between cells for passage of small molecules such as calcium and other ions important for digestive enzyme secretion.

The ducts collect pancreatic enzymes, and the activity of the ductular cells dilutes and alkalinizes pancreatic juice before it is washed out into the small intestine. The duct epithelium is made of cuboidal columnar cells held together by intercellular tight junctions. These ductular cells are packed with mitochondria to supply energy for ion transport. Once stimulated, these cells transport bicarbonate ions into the pancreatic juice as it passes along the duct, with water following in response to the resulting transepithelial osmotic gradient.10

Exocrine Secretions

Pancreatic exocrine secretions, nearly 2.5L/day in volume, can be classified in two groups: organic and inorganic. Organic secretions are proteins such as digestive enzymes and inorganic secretions consist mostly of water and electrolytes. Acinar cells secrete digestive enzymes and ductal cells secrete bicarbonate rich electrolyte solution.11 Depending on the organic secretion, the enzyme component of pancreatic juice is mixed in various proportions with the aqueous component. Greater than 75% of proteins in organic secretions are enzymes and proenzymes; the rest are plasma proteins, trypsin inhibitors, and mucoproteins (Table 1).

Organic Secretions

One of the major purposes of the pancreas is to synthesize digestive enzymes and deliver them to the intestine where they play a critical role in digestion. The four major enzyme groups are proteolytic (eg, chymotrypsin), amylolytic (eg, amylase), lipolytic (eg, lipase), and nuclease digestive enzymes. Some of the enzymes are present in more than one form (e.g., cationic trypsinogen, anionic trypsinogen, and mesotrypsinogen).12 To prevent auto digestion of the pancreas, and hence pancreatitis, enzymes are stored and secreted as inactive precursor forms. Enterokinase, secreted by duodenal mucosa, converts trypsinogen to its active form trypsin, which then catalyzes the activation of the other inactive proenzymes. The acinar cells also secrete a trypsin inhibitor, which inactivates trypsin by disabling this catalytic action.13

Inorganic Secretions

Pancreatic electrolytes (sodium, potassium, chloride, and bicarbonate) mixed with water form an alkaline fluid that is isosmotic with extracellular fluid and helps neutralize gastric acid entering the duodenum.14 Postprandial stimulation mediated mainly by secretin increases secretory flow rate from an average of 0.3 mL/minute in the resting (interdigestive) state to 4.0 mL/minute in the digestive state. The concentrations of bicarbonate and chloride in pancreatic juice change reciprocally as secretory flow rate increases making the osmolality of pancreatic juice independent of flow rate.15

Digestive Enzyme Functions

Amylase The salivary glands and pancreas make amylases. Pancreatic amylase hydrolyzes the 1,4-glycoside linkages of complex carbohydrates and starches. This produces short dextrins, which can then be digested by brush border enzymes like maltose and maltotriose into glucose.

Lipases The majority of dietary lipids in western diets are triglycerides, which cannot be digested by brush border enzymes. Pancreatic triglyceride lipase binds to the oil-water interface of the triglyceride oil droplet where it cleaves the majority of fatty acids from dietary triglycerides.

Bile acids and colipase are important for the full activity of lipases. Bile acids emulsify triglyceride molecules to expand surface area for lipase to act on.16 Colipase forms a complex with lipase and bile salts and anchors lipase to allow it to act in a more hydrophilic environment on the hydrophobic surface of the oil droplet. Carboxyl ester lipase can act on a variety of substrates and is important in digestion of cholesterol esters, lipid-soluble vitamins such as Vitamin A, and triglycerides.

Proteases Pancreatic proteases and gastric pepsin digest all of the complex dietary proteins into short peptides and amino acids for further digestion and absorption in the intestine. The most abundant enzyme is trypsin, which is present in three forms. Cationic trypsinogen, coded by PRSS1 gene, is present in a large proportion, and anionic trypsinogen and mesotrypsinogen, which are coded by PRSS2 and PRSS3 genes, respectively, are present in smaller proportions. All trypsinogens act similarly by attacking the exposed arginine and lysine residues within a peptide chain. Chymotrypsin and elastase are endopeptidases, just like trypsin, that cleave specific peptide bonds adjacent to specific amino acids. These amino acids eventually have greater effects on stimulating pancreatic secretion, inhibiting gastric emptying, regulating small bowel motility, and causing satiety.

Synthesis and Transport of Digestive Enzymes

Protein synthesis occurs in the ribosomes close to the rough endoplasmic reticulum (RER) of acinar cells.17 The cell’s messenger RNA (mRNA) then translates these newly synthesized proteins into exportable proteins. A terminal peptide extension on pancreatic enzymes, known as signal protein, allows attachment and entry of the enzyme into the RER.18 The enzyme and signal protein interact with a membrane protein called a docking protein. This process permits the completion of the translocation, dissociation of the signal protein and mRNA from enzyme, and allows the enzyme to enter RER. Newly synthesized proteins can undergo modifications and conformational changes in the endoplasmic reticulum before being transported to the Golgi complex where further post-translational modification (glycosylation), sorting, and concentration occur.19

Digestive enzymes are then transported to the zymogen granules. A given zymogen granule has various pancreatic proteins mixed in relative proportions depending on their rates of synthesis. The rate of synthesis of a particular enzyme is related to the type of diet. For example, dietary increase in carbohydrates will result in increased expression of amylase compared to other pancreatic enzymes. Zymogen granules then move towards the apical membrane of acinar cells via microtubules and await appropriate neurohormonal stimulus to trigger exocytosis.20

Facilitated by microvilli covering apical surface of acinar cells, exocytosis is a process where the zymogen granule fuses with the apical surface and allows its contents to be released in the ducts.21 This entire process takes less than 1 hour allowing the pancreas to be ready for the next meal by repeating synthesis and packaging of enzymes.

Cellular Regulation of Enzyme Secretion

At the cellular level, secretion of pancreatic juice can be divided into organic and inorganic secretions. Organic secretions containing pancreatic enzymes occurs by regulating acinar cells, and inorganic secretions containing bicarbonate and other electrolytes occurs by regulating ductal cells. Hormonal regulation of acinar and ductal cells is explained in this section and the integrated neurohormonal control of pancreatic secretion is discussed later.

Acinar Cells

Acinar cells express receptors on their basolateral membranes for the following hormones: cholecystokinin (CCK), acetylcholine (ACh), gastrin-releasing peptide (GRP), vasoactive intestinal peptide (VIP), and secretin.22,23 These receptors are divided into two groups based on their mode of stimulating acinar cells (Figure 1). VIP and secretin activate adenylate cyclase, which increases cellular cAMP and facilitates enzyme secretion through cAMP dependent protein kinase A. The other group consisting of acetylcholine, GRP, and CCK lead to an increase in intracellular free calcium concentrations via stimulating cellular metabolism of membrane phosphoinositides.24 This phospholipase C-dependent pathway is the primary stimulus for significant acinar secretion, with cAMP-dependent signaling playing a secondary role.

Ductular Cells

Ductular cells contribute the fluid and bicarbonate components of pancreatic juice. Bicarboanate is predominantly derived from plasma rather than intracellular metabolism. Both the apical and basolateral membranes have polarized epithelial cells and membrane transport proteins that help with ion transportation (Figure 2). Ductal cells are very sensitive to secretin and VIP, both of which increase intracellular cAMP, which in turn leads to opening of CFTR chloride channels initiating secretion.14,15 Bicarbonate enters through the sodium-bicarbonate cotransporter on the basolateral membrane and exits through the CFTR channel on the apical membrane.25 Concomitantly, the sodium-potassium pump keeps the intracellular sodium low thus creating a continual electrochemical force and driving bicarbonate into the ductal cell through the sodium-bicarbonate cotransporter. Water and sodium ions follow paracellularly in response to the electrochemical gradient across the epithelium.

Organ Physiology

Exocrine pancreatic secretion happens during two states: fasting (interdigestive) and after ingestion of a meal (digestive). The interdigestive pattern of secretion begins when the stomach is empty. Secretory activity related to eating (digestive state) occurs in phases: cephalic (20-25%), gastric (10%), and intestinal (approximately 60%-70%). Pancreatic secretion is activated by a combination of neural and hormonal effectors.

Interdigestive Secretions

The interdigestive pancreatic secretions are governed by the cholinergic nervous system, motilin, and pancreatic polypeptide. Secretions follow the cyclical pattern of the migrating myoelectric complex (MMC) [26, 27]. Enzyme secretion occurs every 1 to 2 hours and is associated with the periods of increased motor activity in the stomach and duodenum. In addition to pancreatic enzyme secretion, there is increased secretion of bicarbonate and bile (secondary to partial gallbladder contraction) into the duodenum. The pancreatic secretion during the interdigestive phase is integral to the “housekeeping” function of the MMC to clear the stomach and small intestine of debris including bacteria between meals.26

Digestive Secretions

Secretion with ingestion of a meal is divided into three phases: cephalic, gastric, and intestinal. During the cephalic and gastric phases, secretions are low in volume with high concentrations of digestive enzymes, reflecting stimulation primarily of acinar cells. This stimulation arises from cholinergic vagal input during the cephalic phase, and vago-vagal reflexes activated by gastric distension during the gastric phase. Gastric distention predominantly causes secretion of enzymes with little secretion of water and bicarbonate.28

The intestinal phase begins when chyme leaves stomach and enters the small intestine. During the intestinal phase, ductular secretion is strongly activated, resulting in the production of high volumes of pancreatic juice with decreased concentrations of protein, although the total quantity of enzymes secreted during this phase is actually also markedly increased. Ductular secretion during this phase is driven primarily by the endocrine action of secretin on receptors localized to the basolateral pole of duct epithelial cells. The inputs to the acinar cells during the intestinal phase include CCK as well as neurotransmitters including acetylcholine (ACh), GRP, and VIP.29-31 The large magnitude of the intestinal phase is also attributable to amplification by so-called enteropancreatic reflexes transmitted via the enteric nervous system.

CCK is released from the upper small intestinal mucosa by digestion products of fat, protein, and starch. CCK is a potent stimulus of acinar secretion, acting both directly on CCK-B receptors localized to the basolateral membranes of acinar cells (Figure 1), and via stimulation of vagal afferents close to its site of release in the duodenum, thereby evoking vago-vagal reflexes that stimulate acinar cell secretion via cholinergic and noncholinergic neurotransmitters (the latter including both GRP and VIP).32 In addition to its effects on the pancreas, CCK coordinates the activity of other GI segments and draining organs, including contraction of the gallbladder, relaxation of the sphincter of Oddi, and the slowing of gastric motility to retard gastric emptying and thereby control the rate of delivery of partially digested nutrients to more distal segments of the gut.33 Finally, CCK modulates the activity of secretin in a synergistic fashion by markedly potentiating the effect of secretin as an agonist of pancreatic ductular secretion of bicarbonate.31

The other major regulator of pancreatic secretion is secretin, which is released from S cells in the duodenal mucosa by gastric acid, with a pH threshold of 4.5.34 When the meal enters the small intestine from the stomach, the volume of pancreatic secretions increases rapidly, shifting from a low-volume, protein-rich fluid to a high volume secretion in which enzymes are present at lower concentrations (although in greater absolute amounts, reflecting the effect of CCK and neural effectors on acinar cell secretion). As the secretory rate rises, the pH and bicarbonate concentration in the pancreatic juice rises, with a reciprocal fall in the concentration of chloride ions.34 These latter effects on the composition of the pancreatic juice are mediated predominantly by the endocrine mediator, secretin.35 After the meal, trypsin is free and inhibits intestinal CCK release and pancreatic enzyme secretion. This process is known as feedback inhibition and this effect of trypsin is mediated by intraluminal CCK-releasing factors (CCK-RF) present on the intestinal epithelium.36 In the presence of a meal the digestive proteases are occupied and CCK-RF promotes CCK release and more digestive enzyme secretion. However, after digestion of a meal when there is an excess of digestive proteases in the intestinal lumen, CCK-RF is in turn digested and inactivated so that its ability to augment CCK release and stimulate further pancreatic enzyme secretion ceases (Table 2).

SUMMARY

The pancreas is a complex organ that plays a critical role in the digestion process. Understanding the cellular physiology of acinar and ductal cells lends to grasping the concepts of pancreatic exocrine secretions in various phases of digestion. Knowledge of normal pancreatic secretory functions can help clinicians order appropriate tests, which eventually assist in diagnosing specific pathologies.

Download Tables, Images & References

GASTROINTESTINAL MOTILITY AND FUNCTIONAL BOWEL DISORDERS SERIES #4

Cannabis in Gastrointestinal Disorders

Read Article

For thousands of years, cannabis and its derivatives have been used for the treatment of human diseases including those that present with gastrointestinal (GI) symptoms. Despite the strong evidence supporting the therapeutic role of cannabis in nausea and vomiting related to chemotherapy and cachexia of AIDS, studies on the use of cannabis for other GI disorders are limited and sparse. In this article, we review available clinical evidence in supporting medical cannabis for GI diseases.

Mohammad Bashashati, MD1 Richard W. McCallum, MD2 1Research Associate, Snyder Institute for Chronic Diseases and Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada 2Professor and Founding Chair, Department of Internal Medicine, Texas Tech University Health Sciences Center, El Paso, TX

INTRODUCTION

With over 5000 years of use, cannabis or marijuana, i.e. dried leaves, stems, flowers, and seeds of the plant Cannabis sativa and its compressed resin form, i.e. hashish are the most common consumed illicit drugs worldwide and in the United States.1,2

The chemical derivatives of Cannabis sativa are called “phytocannabinoids”, while there are also synthetic cannabinoids which have cannabinomimetic effects.3

Besides being used recreationally, cannabis and its derivatives have been valuable for the treatment of human diseases for centuries. The term “medical cannabis” or “medical marijuana” describes medicinal use of cannabis or cannabinoids.4 However, based on their potential for being abused, it is very important to define a border between medical and recreational cannabis.5

Under federal law of the United States, cannabis is a Schedule I substance and its use is illegal; however, the District of Columbia and 23 states have legalized medical cannabis. Moreover, in some states recreational use of marijuana is legal and possession of limited amounts of marijuana has been decriminalized. Medical cannabis is approved for chronic, debilitating and long- lasting disorders such as cancer, AIDS and multiple sclerosis and occasionally, this is legally allowed for gastrointestinal (GI) patients with Hepatitis C, Crohn’s disease, nausea and cachexia.6

Due to the legalization of medical cannabis in several states, physicians who see patients with GI diseases may need to prescribe cannabis. Alternatively, they might be asked about the benefits and adverse effects of medical marijuana. In addition, due to decriminalization of cannabis in some states, a practicing physician may see more patients who use cannabis recreationally. Therefore, it is important for a physician to know the therapeutic potentials of cannabis and its natural or synthetic derivatives and the possible adverse effects of cannabis use.

This article highlights the benefits of medical cannabis in GI diseases and discusses possible side effects of medical cannabis and cannabis intoxication.

Cannabis and Cannabinoids

D9-tetrahydrocannabinol (THC) and cannabidiol (CBD) are two major components of cannabis. While THC has psychoactive effects, CBD is devoid of central side effects.7 For medicinal purposes, cannabis can be smoked or vaporized or its extracts/juice can be used orally; however, if taken orally, its absorption is slow and its therapeutic components are less bio-available. More reliable forms of medicinal cannabis in terms of dosage are synthetic compounds which are structurally related to THC, e.g. Nabilone (Cesamet) and Dronabinol (Marinol), as well as accurately measured mixture of THC and CBD extracted from the cannabis plant, e.g. Nabiximols (Sativex). Nabilone and Dronabinol are used orally; however, Nabiximols is an oromucosal spray.8 Besides these, there are synthetic cannabinoid agonists which have not yet been used in the clinic, but have shown significant therapeutic potential in pre-clinical studies as well as the potential for being abused.9

The discovery of endocannabinoid system in 1990s shed light into our understanding of the mechanisms of cannabis function in the body. Moreover, it introduced new potential therapeutic strategies through modulating endocannabinoid system.10 Such an approach is still under clinical evaluation for different clinical conditions and hopefully their results will be released in the early future.

Endocannabinoid System

Basically, our body has endogenous cannabinoids, which are present both in periphery and the central nervous system (CNS). Two main endogenous cannabinoids are anandamide (AEA) and 2-arachydonilglycerol (2-AG). AEA and 2-AG are synthesized on demand from the membrane phospholipids and activate presynaptic cannabinoid receptors 1 and 2 (CB1 and CB2). Endogenous cannabinoids, the enzymes that are involved in the biosynthesis and degradation of them besides CB1 and CB2 receptors are the components of the “endocannabinoid system”.11,12

N-arachidonoylphosphatidylethanolamine phospholipase D (NAPE-PLD) and diacylglycerol lipase (DAGL) are the major enzymes in the biosynthesis of AEA and 2-AG, respectively. On the other hand, AEA is degraded by fatty acid amide hydrolase (FAAH), while monoacylglycerol lipase (MAGL) hydrolyses 2-AG. In addition to these, cyclooxygenase 2 (COX-2) is an important enzyme which metabolizes both AEA and 2-AG and produces prostaglandins.11-13

As mentioned, AEA and 2-AG activate CB1 and CB2 receptors. Another potential receptor for AEA is the transient receptor potential vanilloid type-1 (TRPV1). CB1 and, to some extent, CB2 receptors are present in the central and enteric nervous system. On the other hand, CB2 receptors are more abundant on immune cells, making them a good drug target during inflammation. Despite their variable distribution in different organs, the function of CB1 and CB2 receptors are not distinct and both can either regulate neuronal signaling or might be involved in immune mediated mechanisms.12,14

Endocannabinoid system is present in the GI tract. Although its components have been observed in almost all layers of intestinal sections, they are majorly localized to the enteric nervous system including the myenteric and submucosal plexi. Moreover, there is some evidence to support the presence of CB1 and CB2 receptors on the epithelial cells.10,12 The localization of CB1 and CB2 receptors potentially defines the pharmacophysiology of cannabis and cannabinoids in the GI tract making them a potential target for the treatment of GI disorders such as abdominal pain, diarrhea, nausea and vomiting as well as GI inflammation.

Figure 1 shows the distribution of cannabinoid receptors in the GI tract and CNS regions that control GI function.

The Physiology of Endocannabinoid System in the GI Tract

Our understanding of the physiology of cannabinoids and endocannabinoids relies heavily on findings from the animal studies. In brief, cannabinoids are usually inhibitory in the GI tract through inhibiting vagal pathway.10,12,15 Here, we have summarized the major functions of endocannabinoid system in the GI tract:

  1. Cannabinoids are anti-nociceptive through both CB1 and CB2 receptors. Moreover, enhanced endocannabinoid tone decreases visceral pain in animal models.16,17
  2. Cannabinoids inhibit gastrointestinal motility through CB1 receptors in physiologic conditions and CB1/ CB2 receptors during inflammation.18 Cannabinoids inhibit transient lower esophageal sphincter relaxation,19 delay gastric emptying20 and inhibit intestinal transit or contractility.18
  3. Phytocannabinoids and exogenous cannabinoid agonists inhibit the disordered intestinal permeability during inflammation or after exposure to cholera toxin.21 Moreover, activating CB1 receptors reduces acid secretion in the stomach.22
  4. Cannabinoids induce hyperphagia and increase appetite resulting in weight gain. In addition, cannabinoids including both THC and CBD reduce nausea and vomiting through interacting with CB1, CB2 and TRPV1 receptors as well as through possible interaction with 5-hydroxytryptaminergic (5-HT; serotoninergic) system.12,23
  5. Cannabinoids are anti-inflammatory, and this makes them a potential candidate for the treatment of colitis.24
  6. Cannabinoids induce apoptosis, are antiproliferative and anti-cancerous.25
  7. In the liver, the expression of cannabinoid receptors is low. However, the activation of CB1 receptors is profibrogenic, proinflammatory and pro-steatotic. On the other hand, CB2 receptors inhibit hepatic fibrogenesis.26

Understanding the physiology of endocannabinoid system in the GI tract helps us to define the therapeutic potentials of cannabis and its derivatives in different GI diseases.

Cannabinoids and Gastrointestinal Disorders

While many pre-clinical studies have proved the benefits of cannabinoids and endocannabinoids in GI diseases, clinical studies targeting endocannabinoid system are restricted. This is basically because of the psychoactive effects of these compounds as well as the legal restrictions.

Cannabinoids and Nausea/Vomiting

Conventional antiemetics such as 5-HT3 and NK1 antagonists are helpful in reducing episodes of vomiting; however, they are not usually helpful in reducing the unpleasant sensation of nausea. In contrast, cannabinoids can inhibit both nausea and vomiting. This has been proved in different clinical trials on chemotherapy induced nausea and vomiting (CINV) by using synthetic cannabinoid agonists such as dronabinol, nabilone and levonantradol as well as with THC/CBD mixture (i.e. Sativex). In an earlier study back in 1985, nabilone was compared to prochlorprazine for the treatment of chemotherapy induced emesis. Based on this study, nabilone was superior to prochlorpazine in reducing vomiting episodes.27 Moreover, a recent study that compared dronabinol and ondansetron for delayed CINV showed these medications are similarly effective and their combination is not superior.28 Despite these benefits, due to central side effects, cannabinoid based therapy is not considered the first line treatment in patients with CINV.8,23

Studies on the effects of smoked cannabis and nausea/vomiting are limited. In a study which compared a group of patients who smoked marijuana before/after chemotherapy and another group who had used THC capsule orally, the former group experienced 70-100% symptom relief, while symptom relief in those who had used THC was 76-88%, suggesting a favorable therapeutic role for smoked cannabis in CINV.29 In another study, SÖderpalm et al. tested the effects of smoked cannabis compared to ondansetron in controlling Ipecac induced nausea and vomiting. Smoked cannabis reduced both nausea and vomiting; however, its effects were modest compared to ondansetron.30

A side effect of long-term (generally>5 years) daily marijuana smoking is cannabinoid-induced hyperemesis syndrome, which is particularly observed in male patients and presents with repeated cyclical vomiting and frequent hot bathing. The symptoms usually alleviate after cessation of cannabis smoking. Not all cannabis users develop hyperemesis syndrome. Therefore, the total (cumulative) dose of marijuana, genetic factors, and psychological parameters may contribute to this condition. The pathophysiology of cannabinoid-induced hyperemesis syndrome remains unknown. The few hypotheses which may explain this syndrome are: (a) accumulation of the cannabis derivatives in the brain based on their lipid solubility and long-term half-life, (b) degradation of the cannabis ingredients to some potential emetic metabolites or toxins, (c) delayed gastric emptying induced by cannabis and (d) down-regulation or desensitization of the cannabinoid receptors due to chronic cannabis use.31,32

Cyclic vomiting syndrome (CVS) is a functional GI disorder which presents with nausea and vomiting and epigastric abdominal pain. CVS presents with stereotypical and recurrent episodes of vomiting and abdominal pain with relatively nausea and vomiting free intervals. The pathophysiology of CVS is unknown.32 Studies from our group have shown a high rate of cannabis use in a subset of patients with cyclic vomiting syndrome. These patients needed a higher dose of amitriptyline for the control of their CVS attacks compared to non-cannabis users.33,34 Therefore, it is important to identify cannabinoid induced hyperemesis syndrome as part of the CVS, since a long term goal is decreasing and stopping cannabis use in these patients.

The acute treatment regimen for vomiting attacks in both CVS and cannabinoid-induced hyperemesis syndrome are based on intravenous hydration, intravenous lorazepam due to its sedative effects, anti- emetics and non-narcotic pain medications. Tricyclic antidepressants (TCA) particularly amitriptyline in high doses are recommended for the long-term prophylaxis. Gradual reduction and stopping cannabis use, psychological interventions, anti-anxiety medications and relaxation techniques are necessary in these patients.32

Overall, despite their favorable effects in controlling nausea, cannabinoids are not the first line treatment of nausea and vomiting including CINV as their chronic use may also induce hyperemesis syndrome. Their anti-emetic effects are modest compared to other conventional medications. Moreover, their side effects make them unfavorable treatment for nausea and vomiting. On the other hand, medical cannabis including dronabinol which is available in the United States and has antiemetic effects at 5-10 mg per dose three times a day may be effective in a subset of patients with gastroparesis who present with nausea and vomiting. It is important to recognize that prescribing cannabis based medications is very different from chronic daily cannabis smoking, which can lead to the entity termed cannabinoid-induced hyperemesis syndrome. Whether manipulating endocannabinoid levels is effective in controlling nausea and vomiting is the goal of future studies in this field.

Cannabinoids and Appetite

Cannabis stimulates appetite and increases food intake. Therefore, cannabis-derivatives (e.g. dronabinol) or smoking cannabis are potential treatments for patients with AIDS-associated loss of appetite and cachexia.6,8 On the other hand, to our knowledge, no trial has tested the effects of smoked cannabis in inducing appetite or increasing weight among patients with cancer. Moreover, the data on oral THC as the stimulator of appetite in patients with cancer are not conclusive. Interestingly, recent studies have questioned the general belief of higher rate of obesity in cannabis smokers. Based on two large cohorts including 43,093 and 9282 respondents among US adults aged 18 years or older (2001-2003), the adjusted prevalence of obesity was about 8% lower in participants reporting the use of cannabis at least 3 days per week.35 Therefore, although cannabis may stimulate appetite, its effect on weight gain is not well confirmed. Delayed gastric emptying20 or potential tolerance to chronic cannabis use36 may explain lower weight in chronic cannabis users.

Cannabinoids and IBS

IBS is a functional GI disorder presenting with abdominal pain/discomfort and disturbed bowel habits. The pathophysiology of IBS is not well understood and its treatment is underdeveloped. In IBS endocannabinoid signaling is altered compared to normal population.37 Moreover, cannabinoids can affect both GI motility and visceral sensation.12 Therefore, they are potential candidates for the treatment of IBS especially when it presents with diarrhea.

Studies on the effect of smoked cannabis in IBS are lacking. Based on a clinical trial, dronabinol (5mg) increased colonic compliance and decreased colonic contraction in IBS patients with diarrhea or alternating bowel habits. These responses were varied based on polymorphisms of FAAH and CB1 encoding genes. In this study, dronabinol did not change sensation scores for pain and gas.38 In another study from the same group, treatment with dronabinol did not alter gut transit in IBS-D, but tended to decrease colonic transit in subjects with a specific polymorphism at CB1 encoding gene.39Another study, which tested the effects of dronabinol (up to 10mg) on visceral perception to rectal distension in IBS vs. health controls, showed dronabinol does not affect baseline and stimulated rectal perception.40

Overall the effects of THC as an effective component of cannabis on GI motility in IBS are variable and its effect on abdominal pain and visceral sensation is not promising. Whether manipulating endocannabinoid system is effective in IBS needs further investigations.

Cannabinoids and IBD

Although there is not a well designed case-control study, smoking cannabis looks to be common among patients with inflammatory bowel disease (IBD). Based on the available studies, 39-49% of patients were past/lifetime and 12-14% of them were current marijuana users,41,42 and 18-32% of IBD patients have used marijuana for their IBD symptoms.41,43 Smoking cannabis was more common among patients with Crohn’s disease and a better response to cannabis was observed in these patients.42 On the other hand, long-term cannabis use was correlated with the risk of surgery in patients with Crohn’s disease,43 questioning the therapeutic benefits of cannabis in IBD. Recently, a clinical trial on a small number of patients with Crohn’s disease showed that 8 weeks of smoking cannabis (standardized to 115mg of THC; twice daily) decreased Crohn’s disease activity index (CDAI) more significantly compared to placebo; however, after 2 weeks wash-out period, CDAI was not different in both groups suggesting a temporary but not sustained benefit for smoking cannabis in these patients. C-reactive protein levels did not change after smoking cannabis and endoscopic findings were not collected.44 Although these findings are somehow promising, the temporary effect of cannabis as well as the increased risk of surgery in patients with Crohn’s disease who were long-term users of cannabis questions whether medical cannabis is effective in IBD as advertised in the media and among patients and physicians.

At the moment, we should be cautious in prescribing cannabis for patients with IBD. Larger studies with cannabis or synthetic cannabinoids as well as the modulators of endocannabinoid levels (e.g. FAAH inhibitors) are recommended.

Cannabinoids as Anti-Nocieceptive Agents

As mentioned above, the effect of THC on abdominal pain and visceral sensation in IBS patients is not promising. However, interestingly a recent double- blind controlled trial showed that dronabinol (5mg b.i.d. for 4 weeks) was superior to placebo in increasing pain threshold and decreasing the frequency and intensity of non-reflux related non-cardiac chest pain. No significant adverse effects were noted in this study.45

Therefore, the origin of pain may define the therapeutic efficacy of cannabinoid based medicine in GI diseases with pain. More studies with different regimen are necessary to evaluate the effects of cannabis and its derivatives on visceral sensation and pain.

Cannabinoids and Hepatitis C

Using cannabis as a street drug is common among people with or without hepatitis C. There is a general belief among patients with hepatitis C that cannabis helps with the adverse effects of their medications. This has been shown in studies indicating less adverse effects of interferon and the adherence to HCV treatment in cannabis users or those who were receiving oral cannabinoid-containing medications.46,47 On the other hand, a recent study has shown no association between cannabis use and the likelihood of completing a full course of HCV therapy, interruption of therapy or sustained virological response.48 As the newer and much more effective HCV therapeutic regimens which lack interferon are introduced, the side effects of interferon therapy (e.g. nausea, cachexia, depression) are disappearing and therefore, there will not be a significant indication for the medical cannabis in HCV just based on its potentials in alleviating therapeutic side effects in early future.

Studies on the effects of cannabis and its effect on the progression of liver damage are not conclusive. While some studies have shown that daily cannabis use is associated with liver fibrosis and steatosis in patients with hepatitis C,49-50 others have shown no association between the progression of liver damage and smoking cannabis.48-52

Adverse Effects of Cannabis Use

The disorders related to cannabis are majorly classified as: (1) cannabis intoxication, (2) cannabis use disorder, and (3) cannabis withdrawal.53-55 When recreational cannabis is allowed and medical cannabis is legalized in different states, practicing physicians may see more patients presenting with adverse effects of cannabis use. Many of the adverse effects are self-limited and mild, although psychological interventions, replacement therapies and symptomatic treatment might be necessary.

Cannabis intoxication presents with physiologic and psychiatric presentations of smoking cannabis. Briefly, patients present with anxiety, panic and psychotic symptoms, red eye, tachycardia, dry mouth and increased appetite.53,54

Cannabis use disorder includes daily or near daily heavy or regular cannabis use for a long duration which may present with cannabis dependence syndrome (which is majorly psychological) in around 1 in 10 users, tolerance as what possibly occurs in cannabis hyperemesis syndrome and disturbed personal and social function. Chronic cannabis smoking may cause chronic bronchitis, psychotic symptoms especially in those with other risk factors of these disorders, cognitive impairment in long-term users and after abstinence and impaired educational attainment among adolescents.53,54

Cannabis withdrawal presents with psychiatric symptoms such as insomnia, anxiety, depression, appetite disturbance and physical symptoms of impaired cannabis signaling such as abdominal pain, headache, tremor, and restlessness. Most symptoms usually occur on the first day after abstinence, peak on day 2-6 and are abolished within 14 days.55

Another big concern regarding cannabis abuse is related to synthetic cannabinoids. These drugs, which are predominantly full and potent agonists of CB1 receptors, may induce dangerous side effects such as seizure, hallucination, tachycardia and arrhythmia. They may also affect non-cannabinoid pathways and produce severe intoxication.9 Therefore, we should also be aware of the clinical manifestations of synthetic cannabinoid abuse.

Cannabinoid Receptors Antagonists

Reducing the endocannabiod tone enhances GI motility and decreases the appetite. Therefore, while the purpose of medical cannabis is the induction of endocannabinoid response, reducing endocannabinoid tone with the cannabinoid receptors antagonists is a potential treatment for constipation and obesity. The most famous compound in this area is the CB1 inverse agonist Rimonabant, which was approved in Europe as an anti-obesity medication. This drug was withdrawn from the market based on its serious side effects such as depression and increased suicidal tendency.56

With the invention of peripherally restricted cannabinoid receptors antagonists, we may see other substitutes for Rimonabant with clinical potentials in the future.

RECOMMENDATIONS AND CONCLUSIONS

The pharmacological function of cannabinoid agonists and antagonists may define the potential therapeutic indication of these compounds. These have been summarized in Table 1.

Medical cannabis is effective in patients with chemotherapy induced nausea and vomiting and in cachexia associated with AIDS; however, the evidence to support the role of medical cannabis in other GI diseases is poor. In some cases, cannabis use may associate with the progression of the disease (e.g. hepatitis C) or poor outcome (e.g. risk of surgery in Crohn’s disease). Despite the misguided advertisements for cannabis as a therapy for different GI diseases, we should be cautious in prescribing medical cannabis and should communicate with our patients regarding the adverse effects as well as the limitations of current studies in the field. This may change by larger and well- controlled studies in future.

For patients with CVS who are active cannabis users and patients with cannabinoid-induced hyperemesis syndrome, we recommend gradual decrease and stopping cannabis. During vomiting episodes, lorazepam plus anti-emetics are helpful and for the prophylaxis of vomiting attacks, high-doses of TCAs are recommended.

We believe the compounds that selectively target endocannabinoid metabolism and degradation are the potential medications of future for human diseases including GI disorders.

Download Tables, Images & References

FRONTIERS IN ENDOSCOPY SERIES # 14

Office Management of Hemorrhoid Disease

Read Article

Hemorrhoids account for approximately 3.5 million physician visits and 500 million dollars in costs per annum in the US. The gastroenterologist maintains a unique position between the conservative treatment offered by the primary care physician and the more invasive hemorroidectomy offered by the surgeon. Here we discuss the etiology, classification and diagnosis of hemorrhoids and the procedures for treatment.

INTRODUCTION

Symptomatic hemorrhoids are frequently encountered by gastroenterologists in Western societies. Internal hemorrhoids are the most common cause of recurrent hematochezia in ambulatory adults. Hemorrhoids account for approximately 3.5 million physician visits and 500 million dollars in costs per annum in the USA.1 The gastroenterologist maintains a unique position between the conservative treatment offered by the primary care physician and the more invasive hemorroidectomy offered by the surgeon.

Anatomy

The dentate line is a key anatomical landmark that divides the insensitive rectal mucosa from the richly innervated anal skin. The dentate line dictates classification, sensation, and therapy of hemorrhoids. Internal hemorrhoids are cushions of fibrovascular tissue located proximal to the dentate line. They may prolapse and bleed, but are usually painless. External hemorrhoids are distal to the dentate line and are covered by richly innervated squamous epithelium of the anus. Any therapy whether it be excision or ligation must respect the dentate line if done without anesthesia. Three major cushions reside in the left lateral, right anterior, and right posterior positions. These positions are relatively fixed and allow banding “blind” in certain office techniques. The blood vessels within these cushions are sinusoids that have direct arteriovenous communications between branches of the superior and middle hemorrhoidal arteries, and the superior, rectal continence. Hemorrhoidal cushions contribute to 15-20% of the resting pressure of the anal verge and provide continence by forming complete closure of the anal canal.2

Etiology & Pathophysiology

Risk factors for symptomatic hemorrhoids include low dietary fiber, chronic straining, excessive time on the commode, constipation, diarrhea, pregnancy, and family history.3 The underlying pathophysiology likely involves a multifactorial process involving venous dilation, arterio-venous distention, protrusion of congested anal cushions downward and progressive stretching and collapse of the support structure of the cushions overtime.4 This culminates in prolapse beyond the anal verge. Internal hemorrhoids that remain prolapsed develop ischemia, thrombosis, or gangrene. It is only in this setting that internal hemorrhoids become painful or pruritic. More often encountered is painless bleeding which occurs when the submucosal sinusoids are disrupted. The bleeding is bright red from the pre- sinusoidal arterioles.5

Classification

Hemorrhoids are classified, first, as internal or external relative to the dentate line. (Figure 1) Hemorrhoids above the dentate line are internal hemorrhoids. Hemorrhoids below the dentate line are external hemorrhoids. External hemorrhoids are covered by the very sensitive anoderm. Internal hemorrhoids are further classified by the degree or prolapse, which has direct therapeutic and prognostic consequences:

  • Grade I Do not prolapse below the dentate line; visible only on anoscopy
  • Grade II Prolapse below the dentate line, but spontaneously reduce
  • Grade III Prolapse below the dentate line, but require manual reduction
  • Grade IV Prolapse and stay below the dentate line – not reducible

Nearly all patients with symptomatic Grade I internal hemorrhoids respond well to medical therapy. Grade II and small Grade III internal hemorrhoids respond to non-operative therapy. Large, refractory grade III and grade IV internal hemorrhoids often require surgery.

Clinical Manifestation and Evaluation

Symptoms attributed to hemorrhoids include bleeding, itching and pain.6 Internal hemorrhoids may prolapse and bleed, but are only painful when they have thrombosed, owing to their position proximal to the dentate line. Bleeding is bright red, owing to arterial bleeding from disruption of the arterial-venous connections of the sinusoids. Drops of bright red blood at the end of a bowel movement or on tissue paper, or blood that coats solid brown stool are suggestive features, but are not reliably predictive.7 External hemorrhoids can become very painful when acutely thrombosed.

Three Myths

Three common misconceptions should be dispelled about hemorrhoids. One is the association with portal hypertension. Portal hypertension produces rectal varicies rather than hemorrhoids. In fact, patients with portal hypertension are no more at risk for hemorrhoids.8 The second misconception is that hemorrhoids alone can cause a positive result on stool guaiac tests. This is not true.9 Fecal occult blood should not be attributed to hemorrhoids until the full colon is adequately evaluated. Lastly, the misconception that bleeding from hemorrhoids causes anemia.10 It cannot be overstated that patients who present with anemia require further investigation of the gastrointestinal tract.11 All patients older than 50 years or at high risk for colorectal cancer presenting with rectal bleeding or anorectal symptoms should undergo a full colonoscopy before focusing on internal hemorrhoids as the culprit.

Diagnosis

While many patients with anorectal complaints will ascribe their symptoms to “hemorrhoids”, a careful history and examination will often reveal other anorectal pathology. Careful external inspection of the anus and perirectal area may reveal thrombosed external hemorrhoids, but also anal fissures, fistula, condylomata, rectal abscess, locally advanced rectal cancer, proctitis or pruritus ani. Meanwhile painless bleeding will often require an exam complemented by flexible sigmoidoscopy, colonoscopy or anoscopy. Digital rectal examination is usually unable to detect internal hemorrhoids unless prolapsing or thrombosed. Besides internal hemorrhoids, painless rectal bleeding can also be caused by colorectal carcinoma, polyps, anal fistula, rectal varices, Kaposi sarcoma, and telangiectasias. A painful digital rectal examination should raise the suspicion of concomitant anal fissure which is not uncommon.

Medical Treatment

Conservative treatment begins with the universal recommendation to add fiber to the diet, increased fluid intake, and avoid prolonged straining or sitting on the commode. Prospective studies have shown improvement in pain and bleeding from hemorrhoids with dietary fiber supplementation, though the evidence has not been overwhelming.12,13,14,15 Over the counter sitz baths, suppositories, and topical analgesics and corticosteroids provide symptomatic relief, but no evidence supports their use in reducing actual hemorrhoid swelling, bleeding or prolapse. Long term use of corticosteroid creams is harmful and should be discouraged. Grade I and II hemorrhoids have a decent chance of resolution with medical therapy alone. Taking all comers, the majority of patients presenting with symptomatic hemorrhoids improve with a bowel management program alone.16

Non-Surgical Procedures

When these conservative measures fail, non-surgical procedures are recommended, which is particularly the case with Grade II and small grade III hemorrhoids. It is here where the gastroenterologist can offer an alternative before invasive surgical excision. All of these procedures affix the hemorrhoidal tissue back onto the muscle wall and do not require anesthesia. While they are not excisional, they ablate the mucosa of the hemorrhoid by controlled necrosis via various mechanisms. These include:

  • Sclerotherapy
  • Bipolar Diathermy
  • Direct Current Electrotherapy
  • Infrared Coagulation
  • Rubber Band Ligation

Sclerotherapy is the oldest method and involves injecting a sclerosant (mixture of phenol in oil, quinine, urea and hypertonic saline) into the submucosa at the base of the hemorrhoid through an anoscope using a spinal needle. The technique is suitable for smaller hemorrhoids. Cure rates are reported to be 90%.17 Recurrence rates are 30% at 4 years.18 Significant pain is a limiting factor in this method and has reported in 12-70% of cases.19,20

Bipolar Diathermy, direct current electrotherapy, and infrared coagulation all involve coagulating, occluding and obliterating the hemorrhoidal vascular pedicle proximal to the dentate line. The area of tissue then sloughs and leaves an ulcer that fibrosis at the treatment site. Bipolar diathermy success rates range from 88-100% in randomized trials with first, second and third degree hemorrhoids.21,22,23,24,25 Multiple applications are required at the same site, and 20% of cases require excisional hemorrhoidectomy.26,27 Direct electrotherapy has fallen out of favor because of lengthy treatment times (10 minutes) required to obtain coagulation despite similar success rates (88%) in first, second and third degree hemorrhoids.28,29 A new method using a disposable probe that plugs into any electrosurgical generator is now available. (Figure 1)

Infrared photocoagulation (IRC) produces infrared radiation from a tungsten-halogen lamp which is directed via a polymer probe tip. Office based IRC (Figure 2) and through the scope probes (Figure 3) are available. These are required to make contact with the base of the hemorrhoid under direct vision to deliver 0.5 to 2 second pulses.27 These polses are delivered to all 3 hemorrhjoids in positions at base of the hemorrhoid column (Figure 3a) Success rates range from 67-96% in first and second degree hemorrhoids. Multiple hemorrhoids can be treated at once.30 The swiftness and notably rare incidences of pain and bleeding have made this the most commonly used method amongst the three coagulant therapies.

Rubber band ligation involves completely encircling the redundant mucosa, vascular bundle, and connective tissue with a ligating rubber band. The process must be at least 1-2 cm proximal to the dentate line lest severe immediate pain due to the innervation of the anoderm. The pain can only be relieved by cutting the rubber band, which is a technically challenging feat. Although internal hemorrhoids can be banded using an upper endoscope with a variceal ligation kit, there is a move towards office banding. During colonoscopy or sigmoidoscopy, the internal hemorrhoids are seen during retroflexion but may be missed if air is not suctioned from the rectum (see Figure 4). Figure 2 shows where the band should be placed in relation to the dentate line regardless of the method used. Two commonly used in-offices devices are available. One involves an instrument that physically grasps the mucosa and pulls the tissue into the applicator, and other involves an accessory suction device that sucks the mucosa into the applicator itself. (Figure 5) There are no reported differences in efficacy. Success rates are high and durable, ranging from 80-90% of patients with resolved or improved symptoms on 5 year follow-up.31,32 Complications of rubber band ligation include minor bleeding in less than 5% of cases and severe bleeding in 1-2%. Rare incidences of pelvic sepsis occurring after rubber band ligation have been reported, carrying a high mortality rate (~30%).33,34 Patients should be instructed to seek medical care urgently if symptoms of fever, increased perianal pain, or new onset of urinary retention following the procedure.

Multiple randomized controlled trials have compared each of the above method with one another, but no single has compared all 5 at once. Two meta- analyses concluded that rubber band ligation and infrared coagulation are the most effective. One meta- analysis reported that ligation was more effective because it had more durable benefits, requiring fewer additional treatments for symptomatic recurrence than did coagulation and sclerotherapy.35 However, in a randomized crossover trial comparing rubber band ligation versus infrared coagulation, rubber band ligation resulted in more pain and minor bleeding than infrared coagulation. The use of both ligation and infrared coagulation in combination was 97% successful. Most notably, despite ligation having more pain and bleeding, there was no preference amongst patients for one over the other.36

Surgical Hemorrhoidectomy

Overall, 5-10% of patients with hemorrhoids will require surgery � the vast majority are grade III-IV. Surgical hemorrhoidectomy is indicated in the following scenarios:

  • Failure of Medical and Nonoperative Therapy
  • Symptomatic third-degree, fourth-degree, or significant mixed internal and external hemorrhoids
  • Symptomatic hemorrhoids in the presence of a concomitant anorectal condition that requires surgery
  • Patient preference after discussion of the treatment options with the referring physician and surgeon.37

Surgical hemorrhoidectomy is the most effective and durable treatment overall for third and fourth degree hemorrhoids. Recurrence following a hemorrhoidectomy is quite rare. However, morbidity is higher, especially with respect to incontinence, post- operative pain, urinary retention, bleeding, infection, and anal stenosis. Most patients do not return to work for 2-4 weeks.38,39,40,41,42 The stapled hemorrhoidectomy, developed in 1998, is equally safe as conventional hemorrhoidectomy (open or closed), but with shorter operating time, convalescence and post-operative disability. Effectiveness long term as of yet cannot determined absolutely but hemorrhoid symptoms can recur years later.43

CONCLUSION

Hemorrhoids are normal fibrovascular structures underlying the rectal mucosa and anal skin. Physiologically they maintain the continence of the anus. Symptomatic hemorrhoids bleed, itch or hurt, and require therapy. The dentate line heralds the presence of pain fibers distally, and any intervention must be well proximal to it if done without anesthesia. The degree of prolapse has significant therapeutic implications. For the majority of patients with symptomatic grade I hemorrhoids, lifestyle and dietary changes are effective. Hemorrhoids that fail conservative measures (grade II, III) may be amenable to non-operative procedures that can be done by the gastroenterologist safely in the office setting without anesthesia, including rubber band ligation and infrared photocoagulation. While rubber band ligation seems more durable than photocoagulation, it appears to come at the expense of more pain and minor bleeding. However, patients do not seem to favor one over the other. Surgical hemorrhoidectomy, including a newer stapled technique, is the last line of therapy, and provide the most durable and effective response for refractory hemorrhoids, but expectedly with higher complication rates and prolonged convalescence.

Download Tables, Images & References

jojobethacklinkJojobet GirişJojobet GirişCasibomCasibomiptv satın alluxbetluxbetRulobetbaşakşehir masaj salonukatlaJojobet GirişHoliganbetholiganbetJojobet GirişMarsbahis GirişCasibom GirişMarsbahis Giriş