Search Results

You are looking at 91 - 100 of 331 items for :

Clear All
Restricted access

Paolo Menaspà, Franco M. Impellizzeri, Eric C. Haakonssen, David T. Martin and Chris R. Abbiss

Purpose:

To determine the consistency of commercially available devices used for measuring elevation gain in outdoor activities and sports.

Methods:

Two separate observational validation studies were conducted. Garmin (Forerunner 310XT, Edge 500, Edge 750, and Edge 800; with and without elevation correction) and SRM (Power Control 7) devices were used to measure total elevation gain (TEG) over a 15.7-km mountain climb performed on 6 separate occasions (6 devices; study 1) and during a 138-km cycling event (164 devices; study 2).

Results:

TEG was significantly different between the Garmin and SRM devices (P < .05). The between-devices variability in TEG was lower when measured with the SRM than with the Garmin devices (study 1: 0.2% and 1.5%, respectively). The use of the Garmin elevation-correction option resulted in a 5–10% increase in the TEG.

Conclusions:

While measurements of TEG were relatively consistent within each brand, the measurements differed between the SRM and Garmin devices by as much as 3%. Caution should be taken when comparing elevation-gain data recorded with different settings or with devices of different brands.

Restricted access

Nicola Marsh, Nick Dobbin, Craig Twist and Chris Curtis

This study assessed energy intake and expenditure of international female touch players during an international tournament. Energy intake (food diary) and expenditure (accelerometer, global positioning system) were recorded for 16 female touch players during a four-day tournament, competing in 8.0 ± 1.0 matches; two on Days 1, 2, and 4, and three on Day 3. Total daily energy expenditure (43.6 ± 3.1 Kcal·kg-1 body mass (BM)) was not different (p > .05) from energy intake (39.9 ± 9.4 Kcal·kg-1 BM). Carbohydrate intakes were below current recommendations (6–10 g·kg-1 BM) on Days 1 (4.4 ± 0.6 g·kg-1 BM) and 3 (4.7 ± 1.0 g·kg-1 BM) and significantly below (p < .05) on Day 2 (4.1 ± 1.0 g·kg-1 BM). Protein and fat intakes were consistent with recommendations (protein, 1.2–2.0 g·kg-1 BM: fat, 20–35% total Kcal) across Days 1–3 (protein, 1.9 ± 0.8, 2.2 ± 0.8, and 2.0 ± 0.7 g·kg-1 BM; fat, 35.6 ± 6.8, 38.5 ± 6.4, and 35.9 ± 5.4% total Kcal). Saturated fat intakes were greater (p < .05) than recommendations (10% total Kcal) on Days 1–3 (12.4 ± 2.9, 14.2 ± 5.1, and 12.7 ± 3.5% total Kcal). On average, female touch players maintained energy balance. Carbohydrate intakes appeared insufficient and might have contributed to the reduction (p < .05) in high-intensity running on Day 3. Further research might investigate the applicability of current nutrition recommendations and the role of carbohydrate in multimatch, multiday tournaments.

Restricted access

Dean J. McNamara, Tim J. Gabbett, Geraldine Naughton, Patrick Farhart and Paul Chapman

Purpose:

This study investigated key fatigue and workload variables of cricket fast bowlers and nonfast bowlers during a 7-wk physical-preparation period and 10-d intensified competition period.

Methods:

Twenty-six elite junior cricketers (mean ± SD age 17.7 ± 1.1 y) were classified as fast bowlers (n = 9) or nonfast bowlers (n = 17). Individual workloads were measured via global positioning system technology, and neuromuscular function (countermovement jump [relative power and flight time]), endocrine (salivary testosterone and cortisol concentrations), and perceptual well-being (soreness, mood, stress, sleep quality, and fatigue) markers were recorded.

Results:

Fast bowlers performed greater competition total distance (median [interquartile range] 7049 [3962] m vs 5062 [3694] m), including greater distances at low and high speeds, and more accelerations (40 [32] vs 19 [21]) and had a higher player load (912 [481] arbitrary units vs 697 [424] arbitrary units) than nonfast bowlers. Cortisol concentrations were higher in the physical-preparation (mean ± 90% confidence intervals, % likelihood; d = –0.88 ± 0.39, 100%) and competition phases (d = –0.39 ± 0.30, 85%), and testosterone concentrations, lower (d = 0.56 ± 0.29, 98%), in the competition phase in fast bowlers. Perceptual well-being was poorer in nonfast bowlers during competition only (d = 0.36 ± 0.22, 88%). Differences in neuromuscular function between groups were unclear during physical preparation and competition.

Conclusions:

These findings demonstrate differences in the physical demands of cricket fast bowlers and nonfast bowlers and suggest that these external workloads differentially affect the neuromuscular, endocrine, and perceptual fatigue responses of these players.

Restricted access

Matthias W. Hoppe, Christian Baumgart and Jürgen Freiwald

Purpose:

To investigate differences in running activities between adolescent and adult tennis players during match play. Differences between winning and losing players within each age group were also examined.

Methods:

Forty well-trained male players (20 adolescents, 13 ± 1 y; 20 adults, 25 ± 4 y) played a simulated singles match against an opponent of similar age and ability. Running activities were assessed using portable devices that sampled global positioning system (10 Hz) and inertial-sensor (accelerometer, gyroscope, and magnetometer; 100 Hz) data. Recorded data were examined in terms of velocity, acceleration, deceleration, metabolic power, PlayerLoad, and number of accelerations toward the net and the forehand and backhand corners.

Results:

Adult players spent more time at high velocity (≥4 m/s2), acceleration (≥4 m/s2), deceleration (≤–4 m/s2), and metabolic power (≥20 W/kg) (P ≤ .009, ES = 0.9–1.5) and performed more accelerations (≥2 m/s2) toward the backhand corner (P < .001, ES = 2.6–2.7). No differences between adolescent winning and losing players were evident overall (P ≥ .198, ES = 0.0–0.6). Adult winning players performed more accelerations (2 to <4 m/s2) toward the forehand corner (P = .026, ES = 1.2), whereas adult losing players completed more accelerations (≥2 m/s2) toward the backhand corner (P ≤ .042, ES = 0.9).

Conclusions:

This study shows that differences in running activities between adolescent and adult tennis players exist in high-intensity measures during simulated match play. Furthermore, differences between adolescent and adult players, and also between adult winning and losing players, are present in terms of movement directions. Our findings may be helpful for coaches to design different training drills for both age groups of players.

Restricted access

Ademir F.S. Arruda, Christopher Carling, Vinicius Zanetti, Marcelo S. Aoki, Aaron J. Coutts and Alexandre Moreira

Purpose:

To analyze the effects of a very congested match schedule on the total distance (TD) covered, high-intensity-running (HIR) distance, and frequency of accelerations and body-load impacts (BLIs) performed in a team of under-15 soccer players (N = 10; 15.1 ± 0.2 y, 171.8 ± 4.7 cm, 61 ± 6.0 kg) during an international youth competition.

Methods:

Using global positioning systems, player performances were repeatedly monitored in 5 matches performed over 3 successive days.

Results:

Significant differences were observed between matches (P < .05) for the frequency of accelerations per minute, BLIs, and BLIs per minute. No differences were observed for the TD covered, TD run per minute, number of high-intensity runs, distance covered in HIR, per-minute peak running speed attained, or frequency of accelerations. The frequency of accelerations per minute decreased across the competition while BLIs were higher during the final than in all other matches.

Conclusions:

These results suggest that BLIs and acceleration might be used as an alternative means to represent the external load during congested match schedules rather than measures related to running speed or distance covered.

Restricted access

Cloe Cummins and Rhonda Orr

Objective:

To investigate the impact forces of collision events during both attack and defense in elite rugby league match play and to compare the collision profiles between playing positions.

Participants:

26 elite rugby league players.

Methods:

Player collisions were recorded using an integrated accelerometer in global positioning system units (SPI-Pro X, GPSports). Impact forces of collisions in attack (hit-ups) and defense (tackles) were analyzed from 359 files from outside backs (n = 78), adjustables (n = 97), wide-running forwards (n = 136), and hit-up forwards (n = 48) over 1 National Rugby League season.

Results:

Hit-up forwards were involved in 0.8 collisions/min, significantly more than all other positional groups (wide-running forwards P = .050, adjustables P = .042, and outside backs P = .000). Outside backs experienced 25% fewer collisions per minute than hit-up forwards. Hit-up forwards experienced a collision within the 2 highest classifications of force (≥10 g) every 2.5 min of match play compared with 1 every 5 and 9 min for adjustables and outside backs, respectively. Hit-up forwards performed 0.5 tackles per minute of match play, 5 times that of outside backs (ES = 1.90; 95% CI [0.26,3.16]), and 0.2 hit-ups per minute of match play, twice as many as adjustables.

Conclusions:

During a rugby league match, players are exposed to a significant number of collision events. Positional differences exist, with hit-up and wide-running forwards experiencing greater collision events than adjustables and outside backs. Although these results may be unique to the individual team’s defensive- and attacking-play strategies, they are indicative of the significant collision profiles in professional rugby league.

Open access

Marco Cardinale and Matthew C. Varley

The need to quantify aspects of training to improve training prescription has been the holy grail of sport scientists and coaches for many years. Recently, there has been an increase in scientific interest, possibly due to technological advancements and better equipment to quantify training activities. Over the last few years there has been an increase in the number of studies assessing training load in various athletic cohorts with a bias toward subjective reports and/or quantifications of external load. There is an evident lack of extensive longitudinal studies employing objective internal-load measurements, possibly due to the cost-effectiveness and invasiveness of measures necessary to quantify objective internal loads. Advances in technology might help in developing better wearable tools able to ease the difficulties and costs associated with conducting longitudinal observational studies in athletic cohorts and possibly provide better information on the biological implications of specific external-load patterns. Considering the recent technological developments for monitoring training load and the extensive use of various tools for research and applied work, the aim of this work was to review applications, challenges, and opportunities of various wearable technologies.

Restricted access

Aaron T. Scanlan, Daniel M. Berkelmans, William M. Vickery and Crystal O. Kean

Cricket is a popular international team sport with various game formats ranging from long-duration multiday tests to short-duration Twenty20 game play. The role of batsmen is critical to all game formats, with differing physiological demands imposed during each format. Investigation of the physiological demands imposed during cricket batting has historically been neglected, with much of the research focusing on bowling responses and batting technique. A greater understanding of the physiological demands of the batting role in cricket is required to assist strength and conditioning professionals and coaches with the design of training plans, recovery protocols, and player-management strategies. This brief review provides an updated synthesis of the literature examining the internal (eg, metabolic demands and heart rate) and external (eg, activity work rates) physiological responses to batting in the various game formats, as well as simulated play and small-sided-games training. Although few studies have been done in this area, the summary of data provides important insight regarding physiological responses to batting and highlights that more research on this topic is required. Future research is recommended to combine internal and external measures during actual game play, as well as comparing different game formats and playing levels. In addition, understanding the relationship between batting technique and physiological responses is warranted to gain a more holistic understanding of batting in cricket, as well as to develop appropriate coaching and training strategies.

Restricted access

Tyler L. Goodale, Tim J. Gabbett, Ming-Chang Tsai, Trent Stellingwerff and Jeremy Sheppard

Purpose:

To evaluate the effects of contextual game factors on activity and physiological profiles of international-level women’s rugby sevens players.

Methods:

Twenty international-level female rugby sevens players from the same national team participated in this study. Global positioning system and heart-rate data were collected at 5 World Rugby Women’s Sevens Series events (2013–14 season).

Results:

Total, moderate-speed (0.2–3.5 m/s), and high-speed running (3.5–5.0 m/s) distances were significantly greater in the first half (20.1% ± 4.1%, 17.6% ± 6.9%, 24.5% ± 7.8%), during losses (11.4% ± 6.1%, 6.1% ± 6.4%, 26.9% ± 9.8%), during losses of large magnitudes (≥2 tries) (12.9% ± 8.8%, 6.8% ± 10.0%, 31.2% ± 14.9%), and against top-4 opponents (12.6% ± 8.7%, 11.3% ± 8.5%, 15.5% ± 13.9%). In addition, total distance increased (5.0% ± 5.5%) significantly from day 1 to day 2 of tournaments, and very-high-speed (5.0–6.5 m/s) running distance increased significantly (26.0% ± 14.2%) during losses. Time spent between 90% and 100% of maximum heart rate (16.4% ± 14.5%) and player load (19.0% ± 5.1%) were significantly greater in the second half. No significant differences in physiological or activity profiles were observed between forwards and backs.

Conclusions:

Game half, game outcome, tournament day, opponent rank, and margin of outcome all affected activity profiles, whereas game half affected physiological profiles. No differences in activity or physiological profiles were found between playing positions. Practitioners are advised to develop high-speed running ability in women’s rugby sevens players to prepare them to tolerate the varying factors that affect activity profiles.

Restricted access

Heidi R. Thornton, Jace A. Delaney, Grant M. Duthie and Ben J. Dascombe

Purpose:

To investigate the ability of various internal and external training-load (TL) monitoring measures to predict injury incidence among positional groups in professional rugby league athletes.

Methods:

TL and injury data were collected across 3 seasons (2013–2015) from 25 players competing in National Rugby League competition. Daily TL data were included in the analysis, including session rating of perceived exertion (sRPE-TL), total distance (TD), high-speed-running distance (>5 m/s), and high-metabolic-power distance (HPD; >20 W/kg). Rolling sums were calculated, nontraining days were removed, and athletes’ corresponding injury status was marked as “available” or “unavailable.” Linear (generalized estimating equations) and nonlinear (random forest; RF) statistical methods were adopted.

Results:

Injury risk factors varied according to positional group. For adjustables, the TL variables associated most highly with injury were 7-d TD and 7-d HPD, whereas for hit-up forwards they were sRPE-TL ratio and 14-d TD. For outside backs, 21- and 28-d sRPE-TL were identified, and for wide-running forwards, sRPE-TL ratio. The individual RF models showed that the importance of the TL variables in injury incidence varied between athletes.

Conclusions:

Differences in risk factors were recognized between positional groups and individual athletes, likely due to varied physiological capacities and physical demands. Furthermore, these results suggest that robust machine-learning techniques can appropriately monitor injury risk in professional team-sport athletes.