1 /* Framework for configuring and reading PHY devices
2 * Based on code in sungem_phy.c and gianfar_phy.c
6 * Copyright (c) 2004 Freescale Semiconductor, Inc.
7 * Copyright (c) 2006, 2007 Maciej W. Rozycki
9 * This program is free software; you can redistribute it and/or modify it
10 * under the terms of the GNU General Public License as published by the
11 * Free Software Foundation; either version 2 of the License, or (at your
12 * option) any later version.
16 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18 #include <linux/kernel.h>
19 #include <linux/string.h>
20 #include <linux/errno.h>
21 #include <linux/unistd.h>
22 #include <linux/interrupt.h>
23 #include <linux/delay.h>
24 #include <linux/netdevice.h>
25 #include <linux/etherdevice.h>
26 #include <linux/skbuff.h>
28 #include <linux/module.h>
29 #include <linux/mii.h>
30 #include <linux/ethtool.h>
31 #include <linux/phy.h>
32 #include <linux/timer.h>
33 #include <linux/workqueue.h>
34 #include <linux/mdio.h>
36 #include <linux/uaccess.h>
37 #include <linux/atomic.h>
41 static const char *phy_speed_to_str(int speed
)
57 return "Unsupported (update phy.c)";
61 #define PHY_STATE_STR(_state) \
63 return __stringify(_state); \
65 static const char *phy_state_to_str(enum phy_state st)
69 PHY_STATE_STR(STARTING
)
71 PHY_STATE_STR(PENDING
)
74 PHY_STATE_STR(RUNNING
)
76 PHY_STATE_STR(FORCING
)
77 PHY_STATE_STR(CHANGELINK
)
79 PHY_STATE_STR(RESUMING
)
87 * phy_print_status - Convenience function to print out the current phy status
88 * @phydev: the phy_device struct
90 void phy_print_status(struct phy_device
*phydev
)
93 netdev_info(phydev
->attached_dev
,
94 "Link is Up - %s/%s - flow control %s\n",
95 phy_speed_to_str(phydev
->speed
),
96 DUPLEX_FULL
== phydev
->duplex
? "Full" : "Half",
97 phydev
->pause
? "rx/tx" : "off");
99 netdev_info(phydev
->attached_dev
, "Link is Down\n");
102 EXPORT_SYMBOL(phy_print_status
);
105 * phy_clear_interrupt - Ack the phy device's interrupt
106 * @phydev: the phy_device struct
108 * If the @phydev driver has an ack_interrupt function, call it to
109 * ack and clear the phy device's interrupt.
111 * Returns 0 on success or < 0 on error.
113 static int phy_clear_interrupt(struct phy_device
*phydev
)
115 if (phydev
->drv
->ack_interrupt
)
116 return phydev
->drv
->ack_interrupt(phydev
);
122 * phy_config_interrupt - configure the PHY device for the requested interrupts
123 * @phydev: the phy_device struct
124 * @interrupts: interrupt flags to configure for this @phydev
126 * Returns 0 on success or < 0 on error.
128 static int phy_config_interrupt(struct phy_device
*phydev
, u32 interrupts
)
130 phydev
->interrupts
= interrupts
;
131 if (phydev
->drv
->config_intr
)
132 return phydev
->drv
->config_intr(phydev
);
139 * phy_aneg_done - return auto-negotiation status
140 * @phydev: target phy_device struct
142 * Description: Return the auto-negotiation status from this @phydev
143 * Returns > 0 on success or < 0 on error. 0 means that auto-negotiation
146 int phy_aneg_done(struct phy_device
*phydev
)
148 if (phydev
->drv
->aneg_done
)
149 return phydev
->drv
->aneg_done(phydev
);
151 return genphy_aneg_done(phydev
);
153 EXPORT_SYMBOL(phy_aneg_done
);
155 /* A structure for mapping a particular speed and duplex
156 * combination to a particular SUPPORTED and ADVERTISED value
164 /* A mapping of all SUPPORTED settings to speed/duplex */
165 static const struct phy_setting settings
[] = {
167 .speed
= SPEED_10000
,
168 .duplex
= DUPLEX_FULL
,
169 .setting
= SUPPORTED_10000baseKR_Full
,
172 .speed
= SPEED_10000
,
173 .duplex
= DUPLEX_FULL
,
174 .setting
= SUPPORTED_10000baseKX4_Full
,
177 .speed
= SPEED_10000
,
178 .duplex
= DUPLEX_FULL
,
179 .setting
= SUPPORTED_10000baseT_Full
,
183 .duplex
= DUPLEX_FULL
,
184 .setting
= SUPPORTED_2500baseX_Full
,
188 .duplex
= DUPLEX_FULL
,
189 .setting
= SUPPORTED_1000baseKX_Full
,
193 .duplex
= DUPLEX_FULL
,
194 .setting
= SUPPORTED_1000baseT_Full
,
198 .duplex
= DUPLEX_HALF
,
199 .setting
= SUPPORTED_1000baseT_Half
,
203 .duplex
= DUPLEX_FULL
,
204 .setting
= SUPPORTED_100baseT_Full
,
208 .duplex
= DUPLEX_HALF
,
209 .setting
= SUPPORTED_100baseT_Half
,
213 .duplex
= DUPLEX_FULL
,
214 .setting
= SUPPORTED_10baseT_Full
,
218 .duplex
= DUPLEX_HALF
,
219 .setting
= SUPPORTED_10baseT_Half
,
223 #define MAX_NUM_SETTINGS ARRAY_SIZE(settings)
226 * phy_find_setting - find a PHY settings array entry that matches speed & duplex
227 * @speed: speed to match
228 * @duplex: duplex to match
230 * Description: Searches the settings array for the setting which
231 * matches the desired speed and duplex, and returns the index
232 * of that setting. Returns the index of the last setting if
233 * none of the others match.
235 static inline unsigned int phy_find_setting(int speed
, int duplex
)
237 unsigned int idx
= 0;
239 while (idx
< ARRAY_SIZE(settings
) &&
240 (settings
[idx
].speed
!= speed
|| settings
[idx
].duplex
!= duplex
))
243 return idx
< MAX_NUM_SETTINGS
? idx
: MAX_NUM_SETTINGS
- 1;
247 * phy_find_valid - find a PHY setting that matches the requested features mask
248 * @idx: The first index in settings[] to search
249 * @features: A mask of the valid settings
251 * Description: Returns the index of the first valid setting less
252 * than or equal to the one pointed to by idx, as determined by
253 * the mask in features. Returns the index of the last setting
254 * if nothing else matches.
256 static inline unsigned int phy_find_valid(unsigned int idx
, u32 features
)
258 while (idx
< MAX_NUM_SETTINGS
&& !(settings
[idx
].setting
& features
))
261 return idx
< MAX_NUM_SETTINGS
? idx
: MAX_NUM_SETTINGS
- 1;
265 * phy_supported_speeds - return all speeds currently supported by a phy device
266 * @phy: The phy device to return supported speeds of.
267 * @speeds: buffer to store supported speeds in.
268 * @size: size of speeds buffer.
270 * Description: Returns the number of supported speeds, and fills the speeds
271 * buffer with the supported speeds. If speeds buffer is too small to contain
272 * all currently supported speeds, will return as many speeds as can fit.
274 unsigned int phy_supported_speeds(struct phy_device
*phy
,
275 unsigned int *speeds
,
278 unsigned int count
= 0;
279 unsigned int idx
= 0;
281 while (idx
< MAX_NUM_SETTINGS
&& count
< size
) {
282 idx
= phy_find_valid(idx
, phy
->supported
);
284 if (!(settings
[idx
].setting
& phy
->supported
))
287 /* Assumes settings are grouped by speed */
289 (speeds
[count
- 1] != settings
[idx
].speed
)) {
290 speeds
[count
] = settings
[idx
].speed
;
300 * phy_check_valid - check if there is a valid PHY setting which matches
301 * speed, duplex, and feature mask
302 * @speed: speed to match
303 * @duplex: duplex to match
304 * @features: A mask of the valid settings
306 * Description: Returns true if there is a valid setting, false otherwise.
308 static inline bool phy_check_valid(int speed
, int duplex
, u32 features
)
312 idx
= phy_find_valid(phy_find_setting(speed
, duplex
), features
);
314 return settings
[idx
].speed
== speed
&& settings
[idx
].duplex
== duplex
&&
315 (settings
[idx
].setting
& features
);
319 * phy_sanitize_settings - make sure the PHY is set to supported speed and duplex
320 * @phydev: the target phy_device struct
322 * Description: Make sure the PHY is set to supported speeds and
323 * duplexes. Drop down by one in this order: 1000/FULL,
324 * 1000/HALF, 100/FULL, 100/HALF, 10/FULL, 10/HALF.
326 static void phy_sanitize_settings(struct phy_device
*phydev
)
328 u32 features
= phydev
->supported
;
331 /* Sanitize settings based on PHY capabilities */
332 if ((features
& SUPPORTED_Autoneg
) == 0)
333 phydev
->autoneg
= AUTONEG_DISABLE
;
335 idx
= phy_find_valid(phy_find_setting(phydev
->speed
, phydev
->duplex
),
338 phydev
->speed
= settings
[idx
].speed
;
339 phydev
->duplex
= settings
[idx
].duplex
;
343 * phy_ethtool_sset - generic ethtool sset function, handles all the details
344 * @phydev: target phy_device struct
347 * A few notes about parameter checking:
348 * - We don't set port or transceiver, so we don't care what they
350 * - phy_start_aneg() will make sure forced settings are sane, and
351 * choose the next best ones from the ones selected, so we don't
352 * care if ethtool tries to give us bad values.
354 int phy_ethtool_sset(struct phy_device
*phydev
, struct ethtool_cmd
*cmd
)
356 u32 speed
= ethtool_cmd_speed(cmd
);
358 if (cmd
->phy_address
!= phydev
->mdio
.addr
)
361 /* We make sure that we don't pass unsupported values in to the PHY */
362 cmd
->advertising
&= phydev
->supported
;
364 /* Verify the settings we care about. */
365 if (cmd
->autoneg
!= AUTONEG_ENABLE
&& cmd
->autoneg
!= AUTONEG_DISABLE
)
368 if (cmd
->autoneg
== AUTONEG_ENABLE
&& cmd
->advertising
== 0)
371 if (cmd
->autoneg
== AUTONEG_DISABLE
&&
372 ((speed
!= SPEED_1000
&&
373 speed
!= SPEED_100
&&
374 speed
!= SPEED_10
) ||
375 (cmd
->duplex
!= DUPLEX_HALF
&&
376 cmd
->duplex
!= DUPLEX_FULL
)))
379 phydev
->autoneg
= cmd
->autoneg
;
381 phydev
->speed
= speed
;
383 phydev
->advertising
= cmd
->advertising
;
385 if (AUTONEG_ENABLE
== cmd
->autoneg
)
386 phydev
->advertising
|= ADVERTISED_Autoneg
;
388 phydev
->advertising
&= ~ADVERTISED_Autoneg
;
390 phydev
->duplex
= cmd
->duplex
;
392 phydev
->mdix_ctrl
= cmd
->eth_tp_mdix_ctrl
;
394 /* Restart the PHY */
395 phy_start_aneg(phydev
);
399 EXPORT_SYMBOL(phy_ethtool_sset
);
401 int phy_ethtool_ksettings_set(struct phy_device
*phydev
,
402 const struct ethtool_link_ksettings
*cmd
)
404 u8 autoneg
= cmd
->base
.autoneg
;
405 u8 duplex
= cmd
->base
.duplex
;
406 u32 speed
= cmd
->base
.speed
;
409 if (cmd
->base
.phy_address
!= phydev
->mdio
.addr
)
412 ethtool_convert_link_mode_to_legacy_u32(&advertising
,
413 cmd
->link_modes
.advertising
);
415 /* We make sure that we don't pass unsupported values in to the PHY */
416 advertising
&= phydev
->supported
;
418 /* Verify the settings we care about. */
419 if (autoneg
!= AUTONEG_ENABLE
&& autoneg
!= AUTONEG_DISABLE
)
422 if (autoneg
== AUTONEG_ENABLE
&& advertising
== 0)
425 if (autoneg
== AUTONEG_DISABLE
&&
426 ((speed
!= SPEED_1000
&&
427 speed
!= SPEED_100
&&
428 speed
!= SPEED_10
) ||
429 (duplex
!= DUPLEX_HALF
&&
430 duplex
!= DUPLEX_FULL
)))
433 phydev
->autoneg
= autoneg
;
435 phydev
->speed
= speed
;
437 phydev
->advertising
= advertising
;
439 if (autoneg
== AUTONEG_ENABLE
)
440 phydev
->advertising
|= ADVERTISED_Autoneg
;
442 phydev
->advertising
&= ~ADVERTISED_Autoneg
;
444 phydev
->duplex
= duplex
;
446 phydev
->mdix_ctrl
= cmd
->base
.eth_tp_mdix_ctrl
;
448 /* Restart the PHY */
449 phy_start_aneg(phydev
);
453 EXPORT_SYMBOL(phy_ethtool_ksettings_set
);
455 int phy_ethtool_gset(struct phy_device
*phydev
, struct ethtool_cmd
*cmd
)
457 cmd
->supported
= phydev
->supported
;
459 cmd
->advertising
= phydev
->advertising
;
460 cmd
->lp_advertising
= phydev
->lp_advertising
;
462 ethtool_cmd_speed_set(cmd
, phydev
->speed
);
463 cmd
->duplex
= phydev
->duplex
;
464 if (phydev
->interface
== PHY_INTERFACE_MODE_MOCA
)
465 cmd
->port
= PORT_BNC
;
467 cmd
->port
= PORT_MII
;
468 cmd
->phy_address
= phydev
->mdio
.addr
;
469 cmd
->transceiver
= phy_is_internal(phydev
) ?
470 XCVR_INTERNAL
: XCVR_EXTERNAL
;
471 cmd
->autoneg
= phydev
->autoneg
;
472 cmd
->eth_tp_mdix_ctrl
= phydev
->mdix_ctrl
;
473 cmd
->eth_tp_mdix
= phydev
->mdix
;
477 EXPORT_SYMBOL(phy_ethtool_gset
);
479 int phy_ethtool_ksettings_get(struct phy_device
*phydev
,
480 struct ethtool_link_ksettings
*cmd
)
482 ethtool_convert_legacy_u32_to_link_mode(cmd
->link_modes
.supported
,
485 ethtool_convert_legacy_u32_to_link_mode(cmd
->link_modes
.advertising
,
486 phydev
->advertising
);
488 ethtool_convert_legacy_u32_to_link_mode(cmd
->link_modes
.lp_advertising
,
489 phydev
->lp_advertising
);
491 cmd
->base
.speed
= phydev
->speed
;
492 cmd
->base
.duplex
= phydev
->duplex
;
493 if (phydev
->interface
== PHY_INTERFACE_MODE_MOCA
)
494 cmd
->base
.port
= PORT_BNC
;
496 cmd
->base
.port
= PORT_MII
;
498 cmd
->base
.phy_address
= phydev
->mdio
.addr
;
499 cmd
->base
.autoneg
= phydev
->autoneg
;
500 cmd
->base
.eth_tp_mdix_ctrl
= phydev
->mdix_ctrl
;
501 cmd
->base
.eth_tp_mdix
= phydev
->mdix
;
505 EXPORT_SYMBOL(phy_ethtool_ksettings_get
);
508 * phy_mii_ioctl - generic PHY MII ioctl interface
509 * @phydev: the phy_device struct
510 * @ifr: &struct ifreq for socket ioctl's
511 * @cmd: ioctl cmd to execute
513 * Note that this function is currently incompatible with the
514 * PHYCONTROL layer. It changes registers without regard to
515 * current state. Use at own risk.
517 int phy_mii_ioctl(struct phy_device
*phydev
, struct ifreq
*ifr
, int cmd
)
519 struct mii_ioctl_data
*mii_data
= if_mii(ifr
);
520 u16 val
= mii_data
->val_in
;
521 bool change_autoneg
= false;
525 mii_data
->phy_id
= phydev
->mdio
.addr
;
529 mii_data
->val_out
= mdiobus_read(phydev
->mdio
.bus
,
535 if (mii_data
->phy_id
== phydev
->mdio
.addr
) {
536 switch (mii_data
->reg_num
) {
538 if ((val
& (BMCR_RESET
| BMCR_ANENABLE
)) == 0) {
539 if (phydev
->autoneg
== AUTONEG_ENABLE
)
540 change_autoneg
= true;
541 phydev
->autoneg
= AUTONEG_DISABLE
;
542 if (val
& BMCR_FULLDPLX
)
543 phydev
->duplex
= DUPLEX_FULL
;
545 phydev
->duplex
= DUPLEX_HALF
;
546 if (val
& BMCR_SPEED1000
)
547 phydev
->speed
= SPEED_1000
;
548 else if (val
& BMCR_SPEED100
)
549 phydev
->speed
= SPEED_100
;
550 else phydev
->speed
= SPEED_10
;
553 if (phydev
->autoneg
== AUTONEG_DISABLE
)
554 change_autoneg
= true;
555 phydev
->autoneg
= AUTONEG_ENABLE
;
559 phydev
->advertising
= mii_adv_to_ethtool_adv_t(val
);
560 change_autoneg
= true;
568 mdiobus_write(phydev
->mdio
.bus
, mii_data
->phy_id
,
569 mii_data
->reg_num
, val
);
571 if (mii_data
->phy_id
== phydev
->mdio
.addr
&&
572 mii_data
->reg_num
== MII_BMCR
&&
574 return phy_init_hw(phydev
);
577 return phy_start_aneg(phydev
);
582 if (phydev
->drv
->hwtstamp
)
583 return phydev
->drv
->hwtstamp(phydev
, ifr
);
590 EXPORT_SYMBOL(phy_mii_ioctl
);
593 * phy_start_aneg - start auto-negotiation for this PHY device
594 * @phydev: the phy_device struct
596 * Description: Sanitizes the settings (if we're not autonegotiating
597 * them), and then calls the driver's config_aneg function.
598 * If the PHYCONTROL Layer is operating, we change the state to
599 * reflect the beginning of Auto-negotiation or forcing.
601 int phy_start_aneg(struct phy_device
*phydev
)
605 mutex_lock(&phydev
->lock
);
607 if (AUTONEG_DISABLE
== phydev
->autoneg
)
608 phy_sanitize_settings(phydev
);
610 /* Invalidate LP advertising flags */
611 phydev
->lp_advertising
= 0;
613 err
= phydev
->drv
->config_aneg(phydev
);
617 if (phydev
->state
!= PHY_HALTED
) {
618 if (AUTONEG_ENABLE
== phydev
->autoneg
) {
619 phydev
->state
= PHY_AN
;
620 phydev
->link_timeout
= PHY_AN_TIMEOUT
;
622 phydev
->state
= PHY_FORCING
;
623 phydev
->link_timeout
= PHY_FORCE_TIMEOUT
;
628 mutex_unlock(&phydev
->lock
);
631 EXPORT_SYMBOL(phy_start_aneg
);
634 * phy_start_machine - start PHY state machine tracking
635 * @phydev: the phy_device struct
637 * Description: The PHY infrastructure can run a state machine
638 * which tracks whether the PHY is starting up, negotiating,
639 * etc. This function starts the timer which tracks the state
640 * of the PHY. If you want to maintain your own state machine,
641 * do not call this function.
643 void phy_start_machine(struct phy_device
*phydev
)
645 queue_delayed_work(system_power_efficient_wq
, &phydev
->state_queue
, HZ
);
649 * phy_trigger_machine - trigger the state machine to run
651 * @phydev: the phy_device struct
653 * Description: There has been a change in state which requires that the
654 * state machine runs.
657 static void phy_trigger_machine(struct phy_device
*phydev
)
659 cancel_delayed_work_sync(&phydev
->state_queue
);
660 queue_delayed_work(system_power_efficient_wq
, &phydev
->state_queue
, 0);
664 * phy_stop_machine - stop the PHY state machine tracking
665 * @phydev: target phy_device struct
667 * Description: Stops the state machine timer, sets the state to UP
668 * (unless it wasn't up yet). This function must be called BEFORE
671 void phy_stop_machine(struct phy_device
*phydev
)
673 cancel_delayed_work_sync(&phydev
->state_queue
);
675 mutex_lock(&phydev
->lock
);
676 if (phydev
->state
> PHY_UP
)
677 phydev
->state
= PHY_UP
;
678 mutex_unlock(&phydev
->lock
);
682 * phy_error - enter HALTED state for this PHY device
683 * @phydev: target phy_device struct
685 * Moves the PHY to the HALTED state in response to a read
686 * or write error, and tells the controller the link is down.
687 * Must not be called from interrupt context, or while the
688 * phydev->lock is held.
690 static void phy_error(struct phy_device
*phydev
)
692 mutex_lock(&phydev
->lock
);
693 phydev
->state
= PHY_HALTED
;
694 mutex_unlock(&phydev
->lock
);
696 phy_trigger_machine(phydev
);
700 * phy_interrupt - PHY interrupt handler
701 * @irq: interrupt line
702 * @phy_dat: phy_device pointer
704 * Description: When a PHY interrupt occurs, the handler disables
705 * interrupts, and uses phy_change to handle the interrupt.
707 static irqreturn_t
phy_interrupt(int irq
, void *phy_dat
)
709 struct phy_device
*phydev
= phy_dat
;
711 if (PHY_HALTED
== phydev
->state
)
712 return IRQ_NONE
; /* It can't be ours. */
714 disable_irq_nosync(irq
);
715 atomic_inc(&phydev
->irq_disable
);
723 * phy_enable_interrupts - Enable the interrupts from the PHY side
724 * @phydev: target phy_device struct
726 static int phy_enable_interrupts(struct phy_device
*phydev
)
728 int err
= phy_clear_interrupt(phydev
);
733 return phy_config_interrupt(phydev
, PHY_INTERRUPT_ENABLED
);
737 * phy_disable_interrupts - Disable the PHY interrupts from the PHY side
738 * @phydev: target phy_device struct
740 static int phy_disable_interrupts(struct phy_device
*phydev
)
744 /* Disable PHY interrupts */
745 err
= phy_config_interrupt(phydev
, PHY_INTERRUPT_DISABLED
);
749 /* Clear the interrupt */
750 err
= phy_clear_interrupt(phydev
);
763 * phy_start_interrupts - request and enable interrupts for a PHY device
764 * @phydev: target phy_device struct
766 * Description: Request the interrupt for the given PHY.
767 * If this fails, then we set irq to PHY_POLL.
768 * Otherwise, we enable the interrupts in the PHY.
769 * This should only be called with a valid IRQ number.
770 * Returns 0 on success or < 0 on error.
772 int phy_start_interrupts(struct phy_device
*phydev
)
774 atomic_set(&phydev
->irq_disable
, 0);
775 if (request_threaded_irq(phydev
->irq
, NULL
, phy_interrupt
,
776 IRQF_ONESHOT
| IRQF_SHARED
,
777 phydev_name(phydev
), phydev
) < 0) {
778 pr_warn("%s: Can't get IRQ %d (PHY)\n",
779 phydev
->mdio
.bus
->name
, phydev
->irq
);
780 phydev
->irq
= PHY_POLL
;
784 return phy_enable_interrupts(phydev
);
786 EXPORT_SYMBOL(phy_start_interrupts
);
789 * phy_stop_interrupts - disable interrupts from a PHY device
790 * @phydev: target phy_device struct
792 int phy_stop_interrupts(struct phy_device
*phydev
)
794 int err
= phy_disable_interrupts(phydev
);
799 free_irq(phydev
->irq
, phydev
);
801 /* If work indeed has been cancelled, disable_irq() will have
802 * been left unbalanced from phy_interrupt() and enable_irq()
803 * has to be called so that other devices on the line work.
805 while (atomic_dec_return(&phydev
->irq_disable
) >= 0)
806 enable_irq(phydev
->irq
);
810 EXPORT_SYMBOL(phy_stop_interrupts
);
813 * phy_change - Called by the phy_interrupt to handle PHY changes
814 * @phydev: phy_device struct that interrupted
816 void phy_change(struct phy_device
*phydev
)
818 if (phy_interrupt_is_valid(phydev
)) {
819 if (phydev
->drv
->did_interrupt
&&
820 !phydev
->drv
->did_interrupt(phydev
))
823 if (phy_disable_interrupts(phydev
))
827 mutex_lock(&phydev
->lock
);
828 if ((PHY_RUNNING
== phydev
->state
) || (PHY_NOLINK
== phydev
->state
))
829 phydev
->state
= PHY_CHANGELINK
;
830 mutex_unlock(&phydev
->lock
);
832 if (phy_interrupt_is_valid(phydev
)) {
833 atomic_dec(&phydev
->irq_disable
);
834 enable_irq(phydev
->irq
);
836 /* Reenable interrupts */
837 if (PHY_HALTED
!= phydev
->state
&&
838 phy_config_interrupt(phydev
, PHY_INTERRUPT_ENABLED
))
842 /* reschedule state queue work to run as soon as possible */
843 phy_trigger_machine(phydev
);
847 atomic_dec(&phydev
->irq_disable
);
848 enable_irq(phydev
->irq
);
852 disable_irq(phydev
->irq
);
853 atomic_inc(&phydev
->irq_disable
);
859 * phy_change_work - Scheduled by the phy_mac_interrupt to handle PHY changes
860 * @work: work_struct that describes the work to be done
862 void phy_change_work(struct work_struct
*work
)
864 struct phy_device
*phydev
=
865 container_of(work
, struct phy_device
, phy_queue
);
871 * phy_stop - Bring down the PHY link, and stop checking the status
872 * @phydev: target phy_device struct
874 void phy_stop(struct phy_device
*phydev
)
876 mutex_lock(&phydev
->lock
);
878 if (PHY_HALTED
== phydev
->state
)
881 if (phy_interrupt_is_valid(phydev
)) {
882 /* Disable PHY Interrupts */
883 phy_config_interrupt(phydev
, PHY_INTERRUPT_DISABLED
);
885 /* Clear any pending interrupts */
886 phy_clear_interrupt(phydev
);
889 phydev
->state
= PHY_HALTED
;
892 mutex_unlock(&phydev
->lock
);
894 /* Cannot call flush_scheduled_work() here as desired because
895 * of rtnl_lock(), but PHY_HALTED shall guarantee phy_change()
896 * will not reenable interrupts.
899 EXPORT_SYMBOL(phy_stop
);
902 * phy_start - start or restart a PHY device
903 * @phydev: target phy_device struct
905 * Description: Indicates the attached device's readiness to
906 * handle PHY-related work. Used during startup to start the
907 * PHY, and after a call to phy_stop() to resume operation.
908 * Also used to indicate the MDIO bus has cleared an error
911 void phy_start(struct phy_device
*phydev
)
913 bool do_resume
= false;
916 mutex_lock(&phydev
->lock
);
918 switch (phydev
->state
) {
920 phydev
->state
= PHY_PENDING
;
923 phydev
->state
= PHY_UP
;
926 /* make sure interrupts are re-enabled for the PHY */
927 if (phydev
->irq
!= PHY_POLL
) {
928 err
= phy_enable_interrupts(phydev
);
933 phydev
->state
= PHY_RESUMING
;
939 mutex_unlock(&phydev
->lock
);
941 /* if phy was suspended, bring the physical link up again */
945 phy_trigger_machine(phydev
);
947 EXPORT_SYMBOL(phy_start
);
949 static void phy_adjust_link(struct phy_device
*phydev
)
951 phydev
->adjust_link(phydev
->attached_dev
);
952 phy_led_trigger_change_speed(phydev
);
956 * phy_state_machine - Handle the state machine
957 * @work: work_struct that describes the work to be done
959 void phy_state_machine(struct work_struct
*work
)
961 struct delayed_work
*dwork
= to_delayed_work(work
);
962 struct phy_device
*phydev
=
963 container_of(dwork
, struct phy_device
, state_queue
);
964 bool needs_aneg
= false, do_suspend
= false;
965 enum phy_state old_state
;
969 mutex_lock(&phydev
->lock
);
971 old_state
= phydev
->state
;
973 if (phydev
->drv
->link_change_notify
)
974 phydev
->drv
->link_change_notify(phydev
);
976 switch (phydev
->state
) {
985 phydev
->link_timeout
= PHY_AN_TIMEOUT
;
989 err
= phy_read_status(phydev
);
993 /* If the link is down, give up on negotiation for now */
995 phydev
->state
= PHY_NOLINK
;
996 netif_carrier_off(phydev
->attached_dev
);
997 phy_adjust_link(phydev
);
1001 /* Check if negotiation is done. Break if there's an error */
1002 err
= phy_aneg_done(phydev
);
1006 /* If AN is done, we're running */
1008 phydev
->state
= PHY_RUNNING
;
1009 netif_carrier_on(phydev
->attached_dev
);
1010 phy_adjust_link(phydev
);
1012 } else if (0 == phydev
->link_timeout
--)
1016 if (phy_interrupt_is_valid(phydev
))
1019 err
= phy_read_status(phydev
);
1024 if (AUTONEG_ENABLE
== phydev
->autoneg
) {
1025 err
= phy_aneg_done(phydev
);
1030 phydev
->state
= PHY_AN
;
1031 phydev
->link_timeout
= PHY_AN_TIMEOUT
;
1035 phydev
->state
= PHY_RUNNING
;
1036 netif_carrier_on(phydev
->attached_dev
);
1037 phy_adjust_link(phydev
);
1041 err
= genphy_update_link(phydev
);
1046 phydev
->state
= PHY_RUNNING
;
1047 netif_carrier_on(phydev
->attached_dev
);
1049 if (0 == phydev
->link_timeout
--)
1053 phy_adjust_link(phydev
);
1056 /* Only register a CHANGE if we are polling and link changed
1057 * since latest checking.
1059 if (phydev
->irq
== PHY_POLL
) {
1060 old_link
= phydev
->link
;
1061 err
= phy_read_status(phydev
);
1065 if (old_link
!= phydev
->link
)
1066 phydev
->state
= PHY_CHANGELINK
;
1069 case PHY_CHANGELINK
:
1070 err
= phy_read_status(phydev
);
1075 phydev
->state
= PHY_RUNNING
;
1076 netif_carrier_on(phydev
->attached_dev
);
1078 phydev
->state
= PHY_NOLINK
;
1079 netif_carrier_off(phydev
->attached_dev
);
1082 phy_adjust_link(phydev
);
1084 if (phy_interrupt_is_valid(phydev
))
1085 err
= phy_config_interrupt(phydev
,
1086 PHY_INTERRUPT_ENABLED
);
1091 netif_carrier_off(phydev
->attached_dev
);
1092 phy_adjust_link(phydev
);
1097 if (AUTONEG_ENABLE
== phydev
->autoneg
) {
1098 err
= phy_aneg_done(phydev
);
1102 /* err > 0 if AN is done.
1103 * Otherwise, it's 0, and we're still waiting for AN
1106 err
= phy_read_status(phydev
);
1111 phydev
->state
= PHY_RUNNING
;
1112 netif_carrier_on(phydev
->attached_dev
);
1114 phydev
->state
= PHY_NOLINK
;
1116 phy_adjust_link(phydev
);
1118 phydev
->state
= PHY_AN
;
1119 phydev
->link_timeout
= PHY_AN_TIMEOUT
;
1122 err
= phy_read_status(phydev
);
1127 phydev
->state
= PHY_RUNNING
;
1128 netif_carrier_on(phydev
->attached_dev
);
1130 phydev
->state
= PHY_NOLINK
;
1132 phy_adjust_link(phydev
);
1137 mutex_unlock(&phydev
->lock
);
1140 err
= phy_start_aneg(phydev
);
1141 else if (do_suspend
)
1142 phy_suspend(phydev
);
1147 phydev_dbg(phydev
, "PHY state change %s -> %s\n",
1148 phy_state_to_str(old_state
),
1149 phy_state_to_str(phydev
->state
));
1151 /* Only re-schedule a PHY state machine change if we are polling the
1152 * PHY, if PHY_IGNORE_INTERRUPT is set, then we will be moving
1153 * between states from phy_mac_interrupt()
1155 if (phydev
->irq
== PHY_POLL
)
1156 queue_delayed_work(system_power_efficient_wq
, &phydev
->state_queue
,
1157 PHY_STATE_TIME
* HZ
);
1161 * phy_mac_interrupt - MAC says the link has changed
1162 * @phydev: phy_device struct with changed link
1163 * @new_link: Link is Up/Down.
1165 * Description: The MAC layer is able indicate there has been a change
1166 * in the PHY link status. Set the new link status, and trigger the
1167 * state machine, work a work queue.
1169 void phy_mac_interrupt(struct phy_device
*phydev
, int new_link
)
1171 phydev
->link
= new_link
;
1173 /* Trigger a state machine change */
1174 queue_work(system_power_efficient_wq
, &phydev
->phy_queue
);
1176 EXPORT_SYMBOL(phy_mac_interrupt
);
1178 static inline void mmd_phy_indirect(struct mii_bus
*bus
, int prtad
, int devad
,
1181 /* Write the desired MMD Devad */
1182 bus
->write(bus
, addr
, MII_MMD_CTRL
, devad
);
1184 /* Write the desired MMD register address */
1185 bus
->write(bus
, addr
, MII_MMD_DATA
, prtad
);
1187 /* Select the Function : DATA with no post increment */
1188 bus
->write(bus
, addr
, MII_MMD_CTRL
, (devad
| MII_MMD_CTRL_NOINCR
));
1192 * phy_read_mmd_indirect - reads data from the MMD registers
1193 * @phydev: The PHY device bus
1194 * @prtad: MMD Address
1197 * Description: it reads data from the MMD registers (clause 22 to access to
1198 * clause 45) of the specified phy address.
1199 * To read these register we have:
1200 * 1) Write reg 13 // DEVAD
1201 * 2) Write reg 14 // MMD Address
1202 * 3) Write reg 13 // MMD Data Command for MMD DEVAD
1203 * 3) Read reg 14 // Read MMD data
1205 int phy_read_mmd_indirect(struct phy_device
*phydev
, int prtad
, int devad
)
1207 struct phy_driver
*phydrv
= phydev
->drv
;
1208 int addr
= phydev
->mdio
.addr
;
1211 if (!phydrv
->read_mmd_indirect
) {
1212 struct mii_bus
*bus
= phydev
->mdio
.bus
;
1214 mutex_lock(&bus
->mdio_lock
);
1215 mmd_phy_indirect(bus
, prtad
, devad
, addr
);
1217 /* Read the content of the MMD's selected register */
1218 value
= bus
->read(bus
, addr
, MII_MMD_DATA
);
1219 mutex_unlock(&bus
->mdio_lock
);
1221 value
= phydrv
->read_mmd_indirect(phydev
, prtad
, devad
, addr
);
1225 EXPORT_SYMBOL(phy_read_mmd_indirect
);
1228 * phy_write_mmd_indirect - writes data to the MMD registers
1229 * @phydev: The PHY device
1230 * @prtad: MMD Address
1232 * @data: data to write in the MMD register
1234 * Description: Write data from the MMD registers of the specified
1236 * To write these register we have:
1237 * 1) Write reg 13 // DEVAD
1238 * 2) Write reg 14 // MMD Address
1239 * 3) Write reg 13 // MMD Data Command for MMD DEVAD
1240 * 3) Write reg 14 // Write MMD data
1242 void phy_write_mmd_indirect(struct phy_device
*phydev
, int prtad
,
1243 int devad
, u32 data
)
1245 struct phy_driver
*phydrv
= phydev
->drv
;
1246 int addr
= phydev
->mdio
.addr
;
1248 if (!phydrv
->write_mmd_indirect
) {
1249 struct mii_bus
*bus
= phydev
->mdio
.bus
;
1251 mutex_lock(&bus
->mdio_lock
);
1252 mmd_phy_indirect(bus
, prtad
, devad
, addr
);
1254 /* Write the data into MMD's selected register */
1255 bus
->write(bus
, addr
, MII_MMD_DATA
, data
);
1256 mutex_unlock(&bus
->mdio_lock
);
1258 phydrv
->write_mmd_indirect(phydev
, prtad
, devad
, addr
, data
);
1261 EXPORT_SYMBOL(phy_write_mmd_indirect
);
1264 * phy_init_eee - init and check the EEE feature
1265 * @phydev: target phy_device struct
1266 * @clk_stop_enable: PHY may stop the clock during LPI
1268 * Description: it checks if the Energy-Efficient Ethernet (EEE)
1269 * is supported by looking at the MMD registers 3.20 and 7.60/61
1270 * and it programs the MMD register 3.0 setting the "Clock stop enable"
1273 int phy_init_eee(struct phy_device
*phydev
, bool clk_stop_enable
)
1275 /* According to 802.3az,the EEE is supported only in full duplex-mode.
1276 * Also EEE feature is active when core is operating with MII, GMII
1277 * or RGMII (all kinds). Internal PHYs are also allowed to proceed and
1278 * should return an error if they do not support EEE.
1280 if ((phydev
->duplex
== DUPLEX_FULL
) &&
1281 ((phydev
->interface
== PHY_INTERFACE_MODE_MII
) ||
1282 (phydev
->interface
== PHY_INTERFACE_MODE_GMII
) ||
1283 phy_interface_is_rgmii(phydev
) ||
1284 phy_is_internal(phydev
))) {
1285 int eee_lp
, eee_cap
, eee_adv
;
1289 /* Read phy status to properly get the right settings */
1290 status
= phy_read_status(phydev
);
1294 /* First check if the EEE ability is supported */
1295 eee_cap
= phy_read_mmd_indirect(phydev
, MDIO_PCS_EEE_ABLE
,
1300 cap
= mmd_eee_cap_to_ethtool_sup_t(eee_cap
);
1304 /* Check which link settings negotiated and verify it in
1305 * the EEE advertising registers.
1307 eee_lp
= phy_read_mmd_indirect(phydev
, MDIO_AN_EEE_LPABLE
,
1312 eee_adv
= phy_read_mmd_indirect(phydev
, MDIO_AN_EEE_ADV
,
1317 adv
= mmd_eee_adv_to_ethtool_adv_t(eee_adv
);
1318 lp
= mmd_eee_adv_to_ethtool_adv_t(eee_lp
);
1319 if (!phy_check_valid(phydev
->speed
, phydev
->duplex
, lp
& adv
))
1322 if (clk_stop_enable
) {
1323 /* Configure the PHY to stop receiving xMII
1324 * clock while it is signaling LPI.
1326 int val
= phy_read_mmd_indirect(phydev
, MDIO_CTRL1
,
1331 val
|= MDIO_PCS_CTRL1_CLKSTOP_EN
;
1332 phy_write_mmd_indirect(phydev
, MDIO_CTRL1
,
1336 return 0; /* EEE supported */
1339 return -EPROTONOSUPPORT
;
1341 EXPORT_SYMBOL(phy_init_eee
);
1344 * phy_get_eee_err - report the EEE wake error count
1345 * @phydev: target phy_device struct
1347 * Description: it is to report the number of time where the PHY
1348 * failed to complete its normal wake sequence.
1350 int phy_get_eee_err(struct phy_device
*phydev
)
1352 return phy_read_mmd_indirect(phydev
, MDIO_PCS_EEE_WK_ERR
, MDIO_MMD_PCS
);
1354 EXPORT_SYMBOL(phy_get_eee_err
);
1357 * phy_ethtool_get_eee - get EEE supported and status
1358 * @phydev: target phy_device struct
1359 * @data: ethtool_eee data
1361 * Description: it reportes the Supported/Advertisement/LP Advertisement
1364 int phy_ethtool_get_eee(struct phy_device
*phydev
, struct ethtool_eee
*data
)
1368 /* Get Supported EEE */
1369 val
= phy_read_mmd_indirect(phydev
, MDIO_PCS_EEE_ABLE
, MDIO_MMD_PCS
);
1372 data
->supported
= mmd_eee_cap_to_ethtool_sup_t(val
);
1374 /* Get advertisement EEE */
1375 val
= phy_read_mmd_indirect(phydev
, MDIO_AN_EEE_ADV
, MDIO_MMD_AN
);
1378 data
->advertised
= mmd_eee_adv_to_ethtool_adv_t(val
);
1380 /* Get LP advertisement EEE */
1381 val
= phy_read_mmd_indirect(phydev
, MDIO_AN_EEE_LPABLE
, MDIO_MMD_AN
);
1384 data
->lp_advertised
= mmd_eee_adv_to_ethtool_adv_t(val
);
1388 EXPORT_SYMBOL(phy_ethtool_get_eee
);
1391 * phy_ethtool_set_eee - set EEE supported and status
1392 * @phydev: target phy_device struct
1393 * @data: ethtool_eee data
1395 * Description: it is to program the Advertisement EEE register.
1397 int phy_ethtool_set_eee(struct phy_device
*phydev
, struct ethtool_eee
*data
)
1399 int val
= ethtool_adv_to_mmd_eee_adv_t(data
->advertised
);
1401 /* Mask prohibited EEE modes */
1402 val
&= ~phydev
->eee_broken_modes
;
1404 phy_write_mmd_indirect(phydev
, MDIO_AN_EEE_ADV
, MDIO_MMD_AN
, val
);
1408 EXPORT_SYMBOL(phy_ethtool_set_eee
);
1410 int phy_ethtool_set_wol(struct phy_device
*phydev
, struct ethtool_wolinfo
*wol
)
1412 if (phydev
->drv
->set_wol
)
1413 return phydev
->drv
->set_wol(phydev
, wol
);
1417 EXPORT_SYMBOL(phy_ethtool_set_wol
);
1419 void phy_ethtool_get_wol(struct phy_device
*phydev
, struct ethtool_wolinfo
*wol
)
1421 if (phydev
->drv
->get_wol
)
1422 phydev
->drv
->get_wol(phydev
, wol
);
1424 EXPORT_SYMBOL(phy_ethtool_get_wol
);
1426 int phy_ethtool_get_link_ksettings(struct net_device
*ndev
,
1427 struct ethtool_link_ksettings
*cmd
)
1429 struct phy_device
*phydev
= ndev
->phydev
;
1434 return phy_ethtool_ksettings_get(phydev
, cmd
);
1436 EXPORT_SYMBOL(phy_ethtool_get_link_ksettings
);
1438 int phy_ethtool_set_link_ksettings(struct net_device
*ndev
,
1439 const struct ethtool_link_ksettings
*cmd
)
1441 struct phy_device
*phydev
= ndev
->phydev
;
1446 return phy_ethtool_ksettings_set(phydev
, cmd
);
1448 EXPORT_SYMBOL(phy_ethtool_set_link_ksettings
);
1450 int phy_ethtool_nway_reset(struct net_device
*ndev
)
1452 struct phy_device
*phydev
= ndev
->phydev
;
1457 return genphy_restart_aneg(phydev
);
1459 EXPORT_SYMBOL(phy_ethtool_nway_reset
);