2 * Copyright (c) 2015, Mellanox Technologies. All rights reserved.
4 * This software is available to you under a choice of one of two
5 * licenses. You may choose to be licensed under the terms of the GNU
6 * General Public License (GPL) Version 2, available from the file
7 * COPYING in the main directory of this source tree, or the
8 * OpenIB.org BSD license below:
10 * Redistribution and use in source and binary forms, with or
11 * without modification, are permitted provided that the following
14 * - Redistributions of source code must retain the above
15 * copyright notice, this list of conditions and the following
18 * - Redistributions in binary form must reproduce the above
19 * copyright notice, this list of conditions and the following
20 * disclaimer in the documentation and/or other materials
21 * provided with the distribution.
23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
33 #include <linux/clocksource.h>
37 MLX5E_CYCLES_SHIFT
= 23
40 void mlx5e_fill_hwstamp(struct mlx5e_tstamp
*tstamp
, u64 timestamp
,
41 struct skb_shared_hwtstamps
*hwts
)
45 read_lock(&tstamp
->lock
);
46 nsec
= timecounter_cyc2time(&tstamp
->clock
, timestamp
);
47 read_unlock(&tstamp
->lock
);
49 hwts
->hwtstamp
= ns_to_ktime(nsec
);
52 static u64
mlx5e_read_internal_timer(const struct cyclecounter
*cc
)
54 struct mlx5e_tstamp
*tstamp
= container_of(cc
, struct mlx5e_tstamp
,
57 return mlx5_read_internal_timer(tstamp
->mdev
) & cc
->mask
;
60 static void mlx5e_timestamp_overflow(struct work_struct
*work
)
62 struct delayed_work
*dwork
= to_delayed_work(work
);
63 struct mlx5e_tstamp
*tstamp
= container_of(dwork
, struct mlx5e_tstamp
,
67 write_lock_irqsave(&tstamp
->lock
, flags
);
68 timecounter_read(&tstamp
->clock
);
69 write_unlock_irqrestore(&tstamp
->lock
, flags
);
70 schedule_delayed_work(&tstamp
->overflow_work
, tstamp
->overflow_period
);
73 int mlx5e_hwstamp_set(struct net_device
*dev
, struct ifreq
*ifr
)
75 struct mlx5e_priv
*priv
= netdev_priv(dev
);
76 struct hwtstamp_config config
;
78 if (!MLX5_CAP_GEN(priv
->mdev
, device_frequency_khz
))
81 if (copy_from_user(&config
, ifr
->ifr_data
, sizeof(config
)))
85 switch (config
.tx_type
) {
94 switch (config
.rx_filter
) {
95 case HWTSTAMP_FILTER_NONE
:
96 /* Reset CQE compression to Admin default */
97 mlx5e_modify_rx_cqe_compression(priv
, priv
->params
.rx_cqe_compress_def
);
99 case HWTSTAMP_FILTER_ALL
:
100 case HWTSTAMP_FILTER_SOME
:
101 case HWTSTAMP_FILTER_PTP_V1_L4_EVENT
:
102 case HWTSTAMP_FILTER_PTP_V1_L4_SYNC
:
103 case HWTSTAMP_FILTER_PTP_V1_L4_DELAY_REQ
:
104 case HWTSTAMP_FILTER_PTP_V2_L4_EVENT
:
105 case HWTSTAMP_FILTER_PTP_V2_L4_SYNC
:
106 case HWTSTAMP_FILTER_PTP_V2_L4_DELAY_REQ
:
107 case HWTSTAMP_FILTER_PTP_V2_L2_EVENT
:
108 case HWTSTAMP_FILTER_PTP_V2_L2_SYNC
:
109 case HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ
:
110 case HWTSTAMP_FILTER_PTP_V2_EVENT
:
111 case HWTSTAMP_FILTER_PTP_V2_SYNC
:
112 case HWTSTAMP_FILTER_PTP_V2_DELAY_REQ
:
113 /* Disable CQE compression */
114 netdev_warn(dev
, "Disabling cqe compression");
115 mlx5e_modify_rx_cqe_compression(priv
, false);
116 config
.rx_filter
= HWTSTAMP_FILTER_ALL
;
122 memcpy(&priv
->tstamp
.hwtstamp_config
, &config
, sizeof(config
));
124 return copy_to_user(ifr
->ifr_data
, &config
,
125 sizeof(config
)) ? -EFAULT
: 0;
128 int mlx5e_hwstamp_get(struct net_device
*dev
, struct ifreq
*ifr
)
130 struct mlx5e_priv
*priv
= netdev_priv(dev
);
131 struct hwtstamp_config
*cfg
= &priv
->tstamp
.hwtstamp_config
;
133 if (!MLX5_CAP_GEN(priv
->mdev
, device_frequency_khz
))
136 return copy_to_user(ifr
->ifr_data
, cfg
, sizeof(*cfg
)) ? -EFAULT
: 0;
139 static int mlx5e_ptp_settime(struct ptp_clock_info
*ptp
,
140 const struct timespec64
*ts
)
142 struct mlx5e_tstamp
*tstamp
= container_of(ptp
, struct mlx5e_tstamp
,
144 u64 ns
= timespec64_to_ns(ts
);
147 write_lock_irqsave(&tstamp
->lock
, flags
);
148 timecounter_init(&tstamp
->clock
, &tstamp
->cycles
, ns
);
149 write_unlock_irqrestore(&tstamp
->lock
, flags
);
154 static int mlx5e_ptp_gettime(struct ptp_clock_info
*ptp
,
155 struct timespec64
*ts
)
157 struct mlx5e_tstamp
*tstamp
= container_of(ptp
, struct mlx5e_tstamp
,
162 write_lock_irqsave(&tstamp
->lock
, flags
);
163 ns
= timecounter_read(&tstamp
->clock
);
164 write_unlock_irqrestore(&tstamp
->lock
, flags
);
166 *ts
= ns_to_timespec64(ns
);
171 static int mlx5e_ptp_adjtime(struct ptp_clock_info
*ptp
, s64 delta
)
173 struct mlx5e_tstamp
*tstamp
= container_of(ptp
, struct mlx5e_tstamp
,
177 write_lock_irqsave(&tstamp
->lock
, flags
);
178 timecounter_adjtime(&tstamp
->clock
, delta
);
179 write_unlock_irqrestore(&tstamp
->lock
, flags
);
184 static int mlx5e_ptp_adjfreq(struct ptp_clock_info
*ptp
, s32 delta
)
190 struct mlx5e_tstamp
*tstamp
= container_of(ptp
, struct mlx5e_tstamp
,
198 adj
= tstamp
->nominal_c_mult
;
200 diff
= div_u64(adj
, 1000000000ULL);
202 write_lock_irqsave(&tstamp
->lock
, flags
);
203 timecounter_read(&tstamp
->clock
);
204 tstamp
->cycles
.mult
= neg_adj
? tstamp
->nominal_c_mult
- diff
:
205 tstamp
->nominal_c_mult
+ diff
;
206 write_unlock_irqrestore(&tstamp
->lock
, flags
);
211 static const struct ptp_clock_info mlx5e_ptp_clock_info
= {
212 .owner
= THIS_MODULE
,
213 .max_adj
= 100000000,
219 .adjfreq
= mlx5e_ptp_adjfreq
,
220 .adjtime
= mlx5e_ptp_adjtime
,
221 .gettime64
= mlx5e_ptp_gettime
,
222 .settime64
= mlx5e_ptp_settime
,
226 static void mlx5e_timestamp_init_config(struct mlx5e_tstamp
*tstamp
)
228 tstamp
->hwtstamp_config
.tx_type
= HWTSTAMP_TX_OFF
;
229 tstamp
->hwtstamp_config
.rx_filter
= HWTSTAMP_FILTER_NONE
;
232 void mlx5e_timestamp_init(struct mlx5e_priv
*priv
)
234 struct mlx5e_tstamp
*tstamp
= &priv
->tstamp
;
239 mlx5e_timestamp_init_config(tstamp
);
240 dev_freq
= MLX5_CAP_GEN(priv
->mdev
, device_frequency_khz
);
242 mlx5_core_warn(priv
->mdev
, "invalid device_frequency_khz, aborting HW clock init\n");
245 rwlock_init(&tstamp
->lock
);
246 tstamp
->cycles
.read
= mlx5e_read_internal_timer
;
247 tstamp
->cycles
.shift
= MLX5E_CYCLES_SHIFT
;
248 tstamp
->cycles
.mult
= clocksource_khz2mult(dev_freq
,
249 tstamp
->cycles
.shift
);
250 tstamp
->nominal_c_mult
= tstamp
->cycles
.mult
;
251 tstamp
->cycles
.mask
= CLOCKSOURCE_MASK(41);
252 tstamp
->mdev
= priv
->mdev
;
254 timecounter_init(&tstamp
->clock
, &tstamp
->cycles
,
255 ktime_to_ns(ktime_get_real()));
257 /* Calculate period in seconds to call the overflow watchdog - to make
258 * sure counter is checked at least once every wrap around.
260 ns
= cyclecounter_cyc2ns(&tstamp
->cycles
, tstamp
->cycles
.mask
,
262 do_div(ns
, NSEC_PER_SEC
/ 2 / HZ
);
263 tstamp
->overflow_period
= ns
;
265 INIT_DELAYED_WORK(&tstamp
->overflow_work
, mlx5e_timestamp_overflow
);
266 if (tstamp
->overflow_period
)
267 schedule_delayed_work(&tstamp
->overflow_work
, 0);
269 mlx5_core_warn(priv
->mdev
, "invalid overflow period, overflow_work is not scheduled\n");
271 /* Configure the PHC */
272 tstamp
->ptp_info
= mlx5e_ptp_clock_info
;
273 snprintf(tstamp
->ptp_info
.name
, 16, "mlx5 ptp");
275 tstamp
->ptp
= ptp_clock_register(&tstamp
->ptp_info
,
276 &priv
->mdev
->pdev
->dev
);
277 if (IS_ERR(tstamp
->ptp
)) {
278 mlx5_core_warn(priv
->mdev
, "ptp_clock_register failed %ld\n",
279 PTR_ERR(tstamp
->ptp
));
284 void mlx5e_timestamp_cleanup(struct mlx5e_priv
*priv
)
286 struct mlx5e_tstamp
*tstamp
= &priv
->tstamp
;
288 if (!MLX5_CAP_GEN(priv
->mdev
, device_frequency_khz
))
291 if (priv
->tstamp
.ptp
) {
292 ptp_clock_unregister(priv
->tstamp
.ptp
);
293 priv
->tstamp
.ptp
= NULL
;
296 cancel_delayed_work_sync(&tstamp
->overflow_work
);