Introduction

Polyethylenes (PEs) with elevated crosslinking for improved wear resistance have been in clinical use for more than 10 years. Wear measurements based on clinical radiographs have indicated wear rates substantially lower than for historical PEs [3, 4, 8, 10, 11, 1518, 23, 28, 42], with a corresponding reduction in the incidence and severity of debris-induced osteolysis. The promising clinical performance of the crosslinked PEs has led to their use in hips with larger-diameter balls to provide additional resistance to dislocation [17]. However, a larger-diameter ball has a longer sliding distance per step, which could increase the rate of wear, and there is concern that higher stresses in a thin liner also might substantially accelerate wear.

We, therefore, asked whether the rate of wear of noncrosslinked and crosslinked PE liners would be affected by (1) increasing the ball diameter from 28 mm to 36 mm, (2) the presence of a rigid metal backing, and (3) reducing the thickness of the liner from 6 mm to 3 mm. Using a finite element (FE) model, we also asked (4) whether the contact area and the stresses in the PE were affected by this reduction in thickness.

Materials and Methods

Hip simulator Test 1 (Fig. 1) compared the wear rates of 28- and 36-mm-diameter liners of noncrosslinked and crosslinked PE, with n = 3 for each of the four groups. Three liners of each type of PE were enclosed in polyurethane molds without the metal shells (Fig. 2) and tested for three million cycles to determine the interactive effect of the level of crosslinking and the ball diameter (Fig. 3) Then, to determine the effect of a rigid backing on the wear rate of the liners, they were placed in titanium alloy shells (Duraloc™; DePuy Orthopaedics, Inc, Warsaw, IN), mounted in the simulator using urethane molds (Fig. 2) and tested for an additional 2.5 million cycles. The wear rate for each liner was calculated using linear regression. Statistical analysis was performed using STATA® Version 10 (StataCorp LP, College Station, TX). The level of statistical significance for differences among the mean wear rates was calculated using Student’s t test.

Fig. 1
figure 1

A flow diagram shows Hip Simulator Test 1. I.D. = inner diameter.

Fig. 2
figure 2

On the top row, a photograph shows a PE liner (right) and a urethane mold (left) without metal backing. On the bottom row, the photograph shows a PE liner, metal backing (Duraloc™ type), and urethane mold.

Fig. 3
figure 3

Details of one test station on the orbital-bearing-type hip simulator show a PE liner in a urethane mold, with the femoral ball loaded from above. During the wear test, an acrylic cylinder encloses the liner to contain the serum lubricant.

The liners had a thickness of 6 mm. The noncrosslinked liners were machined from extruded bars of GUR 1050 UHMWPE (Poly-Hi Solidur, Inc, Fort Wayne, IN). The crosslinked liners were machined from extruded bars from the same batch of PE. The bars had been crosslinked by sealing them in foil bags, flushing with inert gas, evacuating, exposing to 5 Mrad of gamma radiation, heating to 155ºC for 24 hours to extinguish residual free radicals, annealing at 120ºC for 24 hours (both processes under partial vacuum), and then slow cooling to room temperature (Marathon™; DePuy). The crosslinked liners were sterilized with gas plasma, rather than gamma radiation, to avoid modifying the degree of crosslinking or reintroducing free radicals. The crystallinity was about 61% for the noncrosslinked PE and about 43% for the crosslinked PE [7].

The femoral balls (Articul/EZE™; DePuy) were fabricated from ASTM F-1537 cobalt-chrome alloy with nominal diameters of 28 mm or 36 mm. The surface roughness of each ball was measured with a Perthometer S8P® profilometer (Mahr Federal, Cincinnati, OH) using a tracing length of 1.75 mm, an evaluation length of 1.25 mm, and a cutoff length of 0.25 mm. The surfaces were implant quality, with an average Ra of about 0.02 μm. The actual diameters of the balls and liners were measured using a Mitutoyo BRT 504 coordinate measuring machine (Mitutoyo America, Aurora, IL) fitted with a Renishaw TP200 touch probe (Renishaw plc, Wolton-under-Edge, UK) and a 2-mm-diameter ruby stylus. The liner diameters averaged 28.20 ± 0.049 mm and 36.29 ± 0.027 mm. The ball diameters averaged 27.93 ± 0.005 mm and 36.0 ± 0.008 mm. For the wear test, the largest ball was paired with the largest liner, giving average radial ball-liner clearances of 132 ± 24 μm for the 28-mm hips and 144 ± 11 μm for the 36-mm hips.

Hip Simulator Test 2 (Fig. 4) evaluated the effect of liner thickness on wear rate by comparing the wear of crosslinked liners all having an inner diameter of 36 mm but with wall thicknesses of 6, 5, 4, or 3 mm, with n = 3 in each group. As in Test 1, the wear rates were calculated using linear regression, and the level of statistical significance of differences in the mean wear rate among the four groups was calculated using Student’s t test. In addition, the relationship between wear rate and liner thickness was calculated using a mixed-model linear regression.

Fig. 4
figure 4

A flow diagram shows Hip Simulator Test 2. I.D. = inner diameter.

All of the liners in Test 2 were fabricated from Marathon™ crosslinked-remelted PE, prepared as described above. After gas plasma sterilization, the liners were artificially aged by heating to 80ºC in air for 21 days and were tested against implant-quality 36-mm-diameter ASTM F-1537 cobalt-chrome balls (DePuy), with an average radial ball-liner clearance of 525 ± 31 μm.

For wear testing, the liners were placed in titanium alloy acetabular shells (Pinnacle™; DePuy), mounted on the simulator using polyurethane molds, and tested for five million cycles. The test lubricant, loading, motion, cleaning intervals, and weighing procedures were the same as for Test 1, including the use of three loaded soak control liners for each of the four liner thicknesses (12 total).

Both wear tests were run on an orbital-bearing-type hip simulator (Shore Western, Monrovia, CA), with the liners mounted below the balls (Fig. 3), under a double-peak Paul-type load cycle [40] with a 2000-N maximum, with motion and loading synchronized at one cycle per second. The lubricant was filter-sterilized bovine serum (HyClone, Logan, UT) diluted with 0.2% sodium azide to retard bacterial degradation and mold growth and with 20 mmol/L EDTA to minimize precipitation of phosphates onto the balls, giving a protein concentration of approximately 90% (63 g/L).

At intervals of 250 thousand cycles, the PE liners and soak controls were rinsed, inspected visually, and replaced on the simulator in fresh serum. At intervals of 0.5 million cycles, the liners were cleaned, vacuum-desiccated, and weighed. The volume of wear was determined by increasing the measured weight loss of the wear liners by the mean weight gain of the soak controls and dividing by the density of UHMWPE (0.93 g/cm3).

To provide further understanding of the relationship between wall thickness and wear of the PE, a FE model was used to calculate the stresses within PE liners having inner diameters of 36 mm and thicknesses of 3, 4, 5, or 6 mm. The radial clearance between the femoral head and the liner was set at 525 μm (ie, comparable to the average clearance in Hip Simulator Test 2) (Fig. 5). Using PATRAN modeling software (v2005; MSC Software Corp, Santa Ana, CA), the liners were represented by 2976 eight-node brick elements and 256 six-node wedge elements. The same element mesh was used for the four liners that had different thicknesses. Since their stiffness was much higher than the PE, the cobalt-chromium head was modeled as an analytical rigid surface, and the titanium alloy shell was represented by a rigid concave spherical surface, using 808 four-node rigid elements. The ball-liner interface was modeled with 808 frictional elements that permitted sliding between the ball and liner, with the coefficient of friction set at 0.065 [31]. The liner-shell interface was modeled as fully bonded. The model had a total of 4848 elements and 14,574 degrees of freedom. The accuracy of the element mesh density was evaluated by increasing the total elements to 76,802. The increase of the stresses due to the higher mesh density was 1%, 4%, and 4% for the maximum contact stress, von Mises stress, and shear stress, respectively, for the liner with a thickness of 4 mm. An elastic modulus of 532 MPa was assigned to the PE elements, obtained from a true stress-strain curve based on the data of uniaxial tensile tests of Marathon™ crosslinked PE provided by the manufacturer. The Poisson’s ratio of the PE elements was set to 0.46. Since it was possible the stress within the PE, particularly in the case of the thin liners, might exceed the yield strength, elements with elastic-plastic deformation characteristics were used, with yield strength of 12.4 MPa and a von Mises yield criterion. The values for the elastic modulus and the yield strength were obtained from tensile tests performed by the manufacturer (DePuy) according to ASTM D638, in which yield is taken as the 0.2% offset strain on a true stress-strain curve (as recommended for PE by Kurtz et al. [27]). Since this material model did not include creep deformation of the PE, which would tend to increase the contact area and, thereby, decrease the stresses, the calculated stresses and the differences in stress among the four thicknesses should be considered worst-case maximum values. A constant force of 2000 N was applied through the center of the ball and normal to the surface of the liner, with the load centered 23° away from the pole of the liner (Fig. 5), corresponding to the maximum load used in the hip simulator tests. The analysis was performed using the ABAQUS program (v6.5; Simulia, Providence, RI).

Fig. 5
figure 5

In a FE model of a PE liner, the load was applied to the inner surface (arrow) through an analytical rigid sphere representing a femoral ball.

Results

In Hip Simulator Test 1 (Table 1), the mean wear rate of the noncrosslinked PE liners without metal shells was 10% greater (p = 0.19) with a 36-mm diameter than with a 28-mm diameter. With metal shells, the mean wear rate was 4% greater (p = 0.4) with a 36-mm diameter than with 28-mm diameter. In contrast, the mean wear rate of the crosslinked PE liners without metal shells was 9% less (p = 0.57) with a 36-mm diameter than with 28-mm diameter. With metal shells, the mean wear rate was 4% greater (p = 0.64) with a 36-mm diameter than with 28-mm diameter. That is, with the crosslinked PE, there was no systematic increase in wear with increasing diameter, and these differences had relatively high p values. The crosslinked cups exhibited a slight gain in weight during the first 0.5 million cycles (Fig. 6). This effect has been reported in other tests of crosslinked PEs [33, 38, 39] and was likely due to the rate of fluid absorption of the wear cups temporarily exceeding that of the control cups. In each of the four test conditions (28- or 36-mm diameter, without or with modular metal backing), the mean wear rate was lower for the 5-Mrad crosslinked PE liners than for the noncrosslinked liners (Table 1), with the reductions ranging from 76% (p < 10−4) for both 28- and 36-mm diameters with metal backing to 83% (p < 10−3) for 36-mm diameter with polyurethane backing.

Table 1 Wear rates of the conventional and 5-Mrad crosslinked polyethylene acetabular cups
Fig. 6
figure 6

A graph shows the volumetric wear of the noncrosslinked (above) and crosslinked (below) PE liners without (left) and with (right) metal backing (Hip Simulator Test 1). Values are presented as mean ± SD. Neither increasing the ball diameter from 28 mm to 36 mm nor introducing the metal backing had a substantial effect on the wear rates.

The mean wear rates of the noncrosslinked PE liners were somewhat lower with metal shells than without them (Table 1), ie, by 8% with the 28-mm diameter (p = 0.07) and by 13% with the 36-mm diameter (p = 0.16). In contrast, the mean wear rates of the crosslinked PE liners were greater with metal shells, ie, by 10% with the 28-mm diameter (p = 0.42) and by 26% with the 36-mm diameter (p = 0.10)

In Hip Simulator Test 2, although there was substantial overlap of the amount of wear for individual liners from the different groups (Fig. 7), the mean wear rate tended to decrease with decreasing liner thickness, such that the mean wear rate of the 3-mm liners was about 19% less (p = 0.17) than that of the 6-mm liners (Table 2).

Fig. 7
figure 7

A graph shows the volumetric wear of the crosslinked PE liners with varying wall thickness (Hip Simulator Test 2). Values are presented as mean ± SD. The mean wear rate tended to decrease as the thickness of the liner decreased from 6 mm to 3 mm.

Table 2 Wear rates of the 5-Mrad crosslinked-remelted polyethylene acetabular cups with varying liner thickness

In the FE modeling of acetabular liner contact stresses, the stresses showed a corresponding increase with the decrease of the thickness of the liners (Fig. 8). The maximum contact stress, the von Mises, and shear stresses were 46%, 17%, and 18% greater in the 3-mm liner than in the 6-mm liner, respectively. While the contact stress for all of the wall thicknesses was greater than the yield strength of 12.4 MPa obtained from the uniaxial tensile tests, the von Mises stress in the liners was lower than the yield strength. Consequently, no yielding was considered to occur with any of the four wall thicknesses.

Fig. 8
figure 8

A graph shows the variation of the maximum contact stress, von Mises stress, and shear stress for liners with different thickness, as calculated with the FE model. Although the maximum value of the contact stress decreased substantially with decreasing thickness, there was much less effect on the maximum values of the von Mises and shear stresses.

In contrast, the contact area showed a corresponding increase with the increase of the thickness of the liners (Fig. 9). The computed contact area between the ball and liner was about 382 mm2 in the 3-mm liner and about 515 mm2 in the 6-mm liner, an increase of 35%.

Fig. 9
figure 9

A graph shows the increase of the contact area between the femoral ball and the PE liner as the thickness of the liner increases, as calculated with the FE model.

Discussion

The improved wear resistance exhibited by the PEs with increased levels of crosslinking has stimulated interest in their use in hip arthroplasties with larger diameters to provide increased resistance to dislocation. However, the sliding distance per step increases in direct proportion to the diameter of the ball, potentially increasing the rate of wear. In addition, for a given outside diameter, a larger ball diameter requires a thinner PE liner. This can result in higher contact and internal stresses in the PE, possibly increasing the risk of excessive wear and/or fracture of the liner. Using a hip simulator, we investigated how the wear rates of noncrosslinked and crosslinked PE liners were affected by (1) increasing the ball diameter from 28 mm to 36-mm; (2) the presence of a rigid metal shell; and (3) reducing the thickness of the liner from 6 mm to 3 mm. We also asked (4) how the contact area and the stresses in the PE were affected by the reduction in thickness.

Our study is associated with several limitations. First a hip simulator cannot exactly reproduce the in vivo wear conditions. Nevertheless, hip simulator tests provide reasonably accurate predictions of the relative wear rates exhibited by different types of PEs in subsequent clinical use [23]. Second, the lower wear observed with the thinner liners occurred under ideal conditions, ie, with the contact zone well within the liner. This has been termed “Mode 1” wear conditions [34], ie, with the acetabular component implanted in the pelvis in an optimal orientation. In clinical practice, acetabular components are sometimes implanted in excessive vertical and/or rotational alignment, which can cause the ball to contact the liner near the rim (a type of Mode 2), or there can be impingement by the neck of the femoral stem (a type of Mode 4). This, in turn, can lead to severe wear and gross fracture of the PE, particularly in liners with external fixation notches that further reduce the thickness of the PE and act as stress concentrations. These types of failure have long been reported for acetabular components featuring conventional PE [14, 20, 35, 43] and, more recently, highly crosslinked PEs [12, 19, 36, 45]. Investigation of these failure modes is beyond the scope of the present study, but they should be taken into account in determining the minimum liner thickness acceptable for clinical use in a particular design of acetabular components.

With historical PEs (either noncrosslinked or moderately crosslinked due to gamma sterilization), the volumetric rate of wear has tended to increase with increasing ball diameter (Table 3) in hip simulator testing [5, 6, 37] and in clinical use [9, 13, 22, 24, 29, 44]. The increase is primarily due to the proportionately longer sliding distance per cycle. This trend also was evident in our study with the noncrosslinked PE, where the mean wear rates of the 36-mm liners were 10% and 4% higher without and with metal backing, respectively. In contrast, with highly crosslinked PEs, there has been little if any increase in the wear rate with increasing diameter in hip simulator tests [21, 37] or clinically [4, 17]. In our study, the mean wear rate of the larger, 36-mm-diameter liner was only slightly higher with the metal backing and was slightly lower without the metal backing (Table 1). The magnitude of the reduction in wear rate exhibited by the 5-Mrad crosslinked PE compared to noncrosslinked PE in our study (Table 1) was comparable to that shown in previous laboratory investigations [32, 33] and in clinical use [28].

Table 3 Comparison of current results with previous studies

The similarity of the wear rates with or without metal backing for either type of PE was consistent with the computational model of Maxian et al. [30], which predicted less than 1% difference in the wear rates of PE liners with or without metal backing.

The observation that the wear rate did not increase as the thickness of the cups decreased was consistent with the results of Kelly et al. [25] who compared the wear rates of acetabular cups of conventional and highly crosslinked PEs in a type of hip simulator closely comparable to that used in our study but with the cups mounted in greater abduction (“near impingement”). The conventional PE cups were gamma sterilized in nitrogen at 3 Mrad and had a 36-mm inner diameter with a 7.9-mm wall thickness (Stryker Orthopaedics, Mahwah, NJ). The PE for the highly crosslinked cups was crosslinked at 9 Mrad using three gamma doses of 3 Mrad, each followed by annealing (X3®; Stryker). These highly crosslinked cups had a 36-mm inner diameter with a 7.9-mm thickness or a 44-mm inner diameter with a 3.8-mm thickness. The mean wear rates were 25, 1.8, and 1.8 mg per million cycles for the conventional cups, 7.9-mm-thick crosslinked cups, and 3.8-mm-thick crosslinked cups, respectively. These authors concluded “crosslinked PE may allow for liners that are thinner than has been traditionally accepted” [25]. However, they also cautioned their results were for an “idealized cup position” and did not take into account the potential effects of edge loading, subluxation, or artificial aging, which has caused fracture of conventional and highly crosslinked liners.

In the FE modeling, the trend for the contact stress to increase as the thickness of the liner decreased (Fig. 8) was consistent with the results obtained by Bartel et al. [2] using analytical solutions and FE analysis. Although the general trend also was in agreement with the results of Plank et al. [41] who evaluated the contact stress of conventional and crosslinked PE using FE analysis and pressure-sensitive film for the comparable situation of a 38-mm ball and a 3-mm-thick PE liner, the contact stress calculated by these authors was 36% lower and von Mises stress was 39% higher than those in our study. This difference might be attributable to a number of factors, such as differences in the material properties, the geometry of the specimens, the clearance between the ball and liner, and the bonding condition assumed between the metal shell and the liners. For example, Bartel et al. [2] found a bonded interface, as used in our study, resulted in a higher contact stress and lower von Mises stress than a debonded interface, as used by Plank et al. [41] and Kurtz et al. [26].

Considering only the increase in the magnitudes of the contact stresses in these studies as the thickness of the liner was reduced, one would predict a corresponding increase in the rate of wear. For example, all other factors equal, Archard’s equation [1] predicts the volumetric rate of wear is directly proportional to the contact stress. Thus, it was surprising the lowest mean rate of wear in our study occurred with the thinnest (3-mm) liners. This trend might have been due, in part, to the contact area being smallest with the 3-mm liners, as indicated by the FE model. In the wear test, this might have been sufficient to more than offset the effect of the increased contact stress, leading to a net reduction in the volumetric rate of wear. In addition, the higher stresses in the thinner liners might have increased the total viscoelastic deformation of the PE, increasing the conformity between the ball and the liner in the contact zone and thereby the potential for at least partial separation of the bearing surfaces by a layer of lubricant, which would tend to reduce the rate of wear.

Taken together, the results of our study indicate, for a 5-Mrad crosslinked PE liner with the acetabular component in proper orientation, the diameter of the ball may be increased and the thickness of the PE decreased within the dimensional limits investigated to improve the resistance to dislocation, with little or no corresponding increase in the volumetric rate of wear and a modest increase in the stresses within the PE.