1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * PTP 1588 clock support
5 * Copyright (C) 2010 OMICRON electronics GmbH
7 #include <linux/device.h>
9 #include <linux/init.h>
10 #include <linux/kernel.h>
11 #include <linux/module.h>
12 #include <linux/posix-clock.h>
13 #include <linux/pps_kernel.h>
14 #include <linux/slab.h>
15 #include <linux/syscalls.h>
16 #include <linux/uaccess.h>
17 #include <linux/debugfs.h>
18 #include <linux/xarray.h>
19 #include <uapi/linux/sched/types.h>
21 #include "ptp_private.h"
23 #define PTP_MAX_ALARMS 4
24 #define PTP_PPS_DEFAULTS (PPS_CAPTUREASSERT | PPS_OFFSETASSERT)
25 #define PTP_PPS_EVENT PPS_CAPTUREASSERT
26 #define PTP_PPS_MODE (PTP_PPS_DEFAULTS | PPS_CANWAIT | PPS_TSFMT_TSPEC)
28 const struct class ptp_class
= {
30 .dev_groups
= ptp_groups
35 static dev_t ptp_devt
;
37 static DEFINE_XARRAY_ALLOC(ptp_clocks_map
);
39 /* time stamp event queue operations */
41 static inline int queue_free(struct timestamp_event_queue
*q
)
43 return PTP_MAX_TIMESTAMPS
- queue_cnt(q
) - 1;
46 static void enqueue_external_timestamp(struct timestamp_event_queue
*queue
,
47 struct ptp_clock_event
*src
)
49 struct ptp_extts_event
*dst
;
50 struct timespec64 offset_ts
;
55 if (src
->type
== PTP_CLOCK_EXTTS
) {
56 seconds
= div_u64_rem(src
->timestamp
, 1000000000, &remainder
);
57 } else if (src
->type
== PTP_CLOCK_EXTOFF
) {
58 offset_ts
= ns_to_timespec64(src
->offset
);
59 seconds
= offset_ts
.tv_sec
;
60 remainder
= offset_ts
.tv_nsec
;
62 WARN(1, "%s: unknown type %d\n", __func__
, src
->type
);
66 spin_lock_irqsave(&queue
->lock
, flags
);
68 dst
= &queue
->buf
[queue
->tail
];
69 dst
->index
= src
->index
;
70 dst
->flags
= PTP_EXTTS_EVENT_VALID
;
72 dst
->t
.nsec
= remainder
;
73 if (src
->type
== PTP_CLOCK_EXTOFF
)
74 dst
->flags
|= PTP_EXT_OFFSET
;
76 /* Both WRITE_ONCE() are paired with READ_ONCE() in queue_cnt() */
77 if (!queue_free(queue
))
78 WRITE_ONCE(queue
->head
, (queue
->head
+ 1) % PTP_MAX_TIMESTAMPS
);
80 WRITE_ONCE(queue
->tail
, (queue
->tail
+ 1) % PTP_MAX_TIMESTAMPS
);
82 spin_unlock_irqrestore(&queue
->lock
, flags
);
85 /* posix clock implementation */
87 static int ptp_clock_getres(struct posix_clock
*pc
, struct timespec64
*tp
)
94 static int ptp_clock_settime(struct posix_clock
*pc
, const struct timespec64
*tp
)
96 struct ptp_clock
*ptp
= container_of(pc
, struct ptp_clock
, clock
);
98 if (ptp_clock_freerun(ptp
)) {
99 pr_err("ptp: physical clock is free running\n");
103 return ptp
->info
->settime64(ptp
->info
, tp
);
106 static int ptp_clock_gettime(struct posix_clock
*pc
, struct timespec64
*tp
)
108 struct ptp_clock
*ptp
= container_of(pc
, struct ptp_clock
, clock
);
111 if (ptp
->info
->gettimex64
)
112 err
= ptp
->info
->gettimex64(ptp
->info
, tp
, NULL
);
114 err
= ptp
->info
->gettime64(ptp
->info
, tp
);
118 static int ptp_clock_adjtime(struct posix_clock
*pc
, struct __kernel_timex
*tx
)
120 struct ptp_clock
*ptp
= container_of(pc
, struct ptp_clock
, clock
);
121 struct ptp_clock_info
*ops
;
122 int err
= -EOPNOTSUPP
;
124 if (ptp_clock_freerun(ptp
)) {
125 pr_err("ptp: physical clock is free running\n");
131 if (tx
->modes
& ADJ_SETOFFSET
) {
132 struct timespec64 ts
;
136 ts
.tv_sec
= tx
->time
.tv_sec
;
137 ts
.tv_nsec
= tx
->time
.tv_usec
;
139 if (!(tx
->modes
& ADJ_NANO
))
142 if ((unsigned long) ts
.tv_nsec
>= NSEC_PER_SEC
)
145 kt
= timespec64_to_ktime(ts
);
146 delta
= ktime_to_ns(kt
);
147 err
= ops
->adjtime(ops
, delta
);
148 } else if (tx
->modes
& ADJ_FREQUENCY
) {
149 long ppb
= scaled_ppm_to_ppb(tx
->freq
);
150 if (ppb
> ops
->max_adj
|| ppb
< -ops
->max_adj
)
152 err
= ops
->adjfine(ops
, tx
->freq
);
153 ptp
->dialed_frequency
= tx
->freq
;
154 } else if (tx
->modes
& ADJ_OFFSET
) {
156 s32 max_phase_adj
= ops
->getmaxphase(ops
);
157 s32 offset
= tx
->offset
;
159 if (!(tx
->modes
& ADJ_NANO
))
160 offset
*= NSEC_PER_USEC
;
162 if (offset
> max_phase_adj
|| offset
< -max_phase_adj
)
165 err
= ops
->adjphase(ops
, offset
);
167 } else if (tx
->modes
== 0) {
168 tx
->freq
= ptp
->dialed_frequency
;
175 static struct posix_clock_operations ptp_clock_ops
= {
176 .owner
= THIS_MODULE
,
177 .clock_adjtime
= ptp_clock_adjtime
,
178 .clock_gettime
= ptp_clock_gettime
,
179 .clock_getres
= ptp_clock_getres
,
180 .clock_settime
= ptp_clock_settime
,
183 .release
= ptp_release
,
188 static void ptp_clock_release(struct device
*dev
)
190 struct ptp_clock
*ptp
= container_of(dev
, struct ptp_clock
, dev
);
191 struct timestamp_event_queue
*tsevq
;
194 ptp_cleanup_pin_groups(ptp
);
195 kfree(ptp
->vclock_index
);
196 mutex_destroy(&ptp
->pincfg_mux
);
197 mutex_destroy(&ptp
->n_vclocks_mux
);
198 /* Delete first entry */
199 spin_lock_irqsave(&ptp
->tsevqs_lock
, flags
);
200 tsevq
= list_first_entry(&ptp
->tsevqs
, struct timestamp_event_queue
,
202 list_del(&tsevq
->qlist
);
203 spin_unlock_irqrestore(&ptp
->tsevqs_lock
, flags
);
204 bitmap_free(tsevq
->mask
);
206 debugfs_remove(ptp
->debugfs_root
);
207 xa_erase(&ptp_clocks_map
, ptp
->index
);
211 static int ptp_getcycles64(struct ptp_clock_info
*info
, struct timespec64
*ts
)
213 if (info
->getcyclesx64
)
214 return info
->getcyclesx64(info
, ts
, NULL
);
216 return info
->gettime64(info
, ts
);
219 static void ptp_aux_kworker(struct kthread_work
*work
)
221 struct ptp_clock
*ptp
= container_of(work
, struct ptp_clock
,
223 struct ptp_clock_info
*info
= ptp
->info
;
226 delay
= info
->do_aux_work(info
);
229 kthread_queue_delayed_work(ptp
->kworker
, &ptp
->aux_work
, delay
);
232 /* public interface */
234 struct ptp_clock
*ptp_clock_register(struct ptp_clock_info
*info
,
235 struct device
*parent
)
237 struct ptp_clock
*ptp
;
238 struct timestamp_event_queue
*queue
= NULL
;
239 int err
, index
, major
= MAJOR(ptp_devt
);
240 char debugfsname
[16];
243 if (info
->n_alarm
> PTP_MAX_ALARMS
)
244 return ERR_PTR(-EINVAL
);
246 /* Initialize a clock structure. */
247 ptp
= kzalloc(sizeof(struct ptp_clock
), GFP_KERNEL
);
253 err
= xa_alloc(&ptp_clocks_map
, &index
, ptp
, xa_limit_31b
,
258 ptp
->clock
.ops
= ptp_clock_ops
;
260 ptp
->devid
= MKDEV(major
, index
);
262 INIT_LIST_HEAD(&ptp
->tsevqs
);
263 queue
= kzalloc(sizeof(*queue
), GFP_KERNEL
);
266 goto no_memory_queue
;
268 list_add_tail(&queue
->qlist
, &ptp
->tsevqs
);
269 spin_lock_init(&ptp
->tsevqs_lock
);
270 queue
->mask
= bitmap_alloc(PTP_MAX_CHANNELS
, GFP_KERNEL
);
273 goto no_memory_bitmap
;
275 bitmap_set(queue
->mask
, 0, PTP_MAX_CHANNELS
);
276 spin_lock_init(&queue
->lock
);
277 mutex_init(&ptp
->pincfg_mux
);
278 mutex_init(&ptp
->n_vclocks_mux
);
279 init_waitqueue_head(&ptp
->tsev_wq
);
281 if (ptp
->info
->getcycles64
|| ptp
->info
->getcyclesx64
) {
282 ptp
->has_cycles
= true;
283 if (!ptp
->info
->getcycles64
&& ptp
->info
->getcyclesx64
)
284 ptp
->info
->getcycles64
= ptp_getcycles64
;
286 /* Free running cycle counter not supported, use time. */
287 ptp
->info
->getcycles64
= ptp_getcycles64
;
289 if (ptp
->info
->gettimex64
)
290 ptp
->info
->getcyclesx64
= ptp
->info
->gettimex64
;
292 if (ptp
->info
->getcrosststamp
)
293 ptp
->info
->getcrosscycles
= ptp
->info
->getcrosststamp
;
296 if (ptp
->info
->do_aux_work
) {
297 kthread_init_delayed_work(&ptp
->aux_work
, ptp_aux_kworker
);
298 ptp
->kworker
= kthread_create_worker(0, "ptp%d", ptp
->index
);
299 if (IS_ERR(ptp
->kworker
)) {
300 err
= PTR_ERR(ptp
->kworker
);
301 pr_err("failed to create ptp aux_worker %d\n", err
);
306 /* PTP virtual clock is being registered under physical clock */
307 if (parent
&& parent
->class && parent
->class->name
&&
308 strcmp(parent
->class->name
, "ptp") == 0)
309 ptp
->is_virtual_clock
= true;
311 if (!ptp
->is_virtual_clock
) {
312 ptp
->max_vclocks
= PTP_DEFAULT_MAX_VCLOCKS
;
314 size
= sizeof(int) * ptp
->max_vclocks
;
315 ptp
->vclock_index
= kzalloc(size
, GFP_KERNEL
);
316 if (!ptp
->vclock_index
) {
318 goto no_mem_for_vclocks
;
322 err
= ptp_populate_pin_groups(ptp
);
326 /* Register a new PPS source. */
328 struct pps_source_info pps
;
329 memset(&pps
, 0, sizeof(pps
));
330 snprintf(pps
.name
, PPS_MAX_NAME_LEN
, "ptp%d", index
);
331 pps
.mode
= PTP_PPS_MODE
;
332 pps
.owner
= info
->owner
;
333 ptp
->pps_source
= pps_register_source(&pps
, PTP_PPS_DEFAULTS
);
334 if (IS_ERR(ptp
->pps_source
)) {
335 err
= PTR_ERR(ptp
->pps_source
);
336 pr_err("failed to register pps source\n");
339 ptp
->pps_source
->lookup_cookie
= ptp
;
342 /* Initialize a new device of our class in our clock structure. */
343 device_initialize(&ptp
->dev
);
344 ptp
->dev
.devt
= ptp
->devid
;
345 ptp
->dev
.class = &ptp_class
;
346 ptp
->dev
.parent
= parent
;
347 ptp
->dev
.groups
= ptp
->pin_attr_groups
;
348 ptp
->dev
.release
= ptp_clock_release
;
349 dev_set_drvdata(&ptp
->dev
, ptp
);
350 dev_set_name(&ptp
->dev
, "ptp%d", ptp
->index
);
352 /* Create a posix clock and link it to the device. */
353 err
= posix_clock_register(&ptp
->clock
, &ptp
->dev
);
356 pps_unregister_source(ptp
->pps_source
);
359 kthread_destroy_worker(ptp
->kworker
);
361 put_device(&ptp
->dev
);
363 pr_err("failed to create posix clock\n");
367 /* Debugfs initialization */
368 snprintf(debugfsname
, sizeof(debugfsname
), "ptp%d", ptp
->index
);
369 ptp
->debugfs_root
= debugfs_create_dir(debugfsname
, NULL
);
374 ptp_cleanup_pin_groups(ptp
);
376 kfree(ptp
->vclock_index
);
379 kthread_destroy_worker(ptp
->kworker
);
381 mutex_destroy(&ptp
->pincfg_mux
);
382 mutex_destroy(&ptp
->n_vclocks_mux
);
383 bitmap_free(queue
->mask
);
385 list_del(&queue
->qlist
);
388 xa_erase(&ptp_clocks_map
, index
);
394 EXPORT_SYMBOL(ptp_clock_register
);
396 static int unregister_vclock(struct device
*dev
, void *data
)
398 struct ptp_clock
*ptp
= dev_get_drvdata(dev
);
400 ptp_vclock_unregister(info_to_vclock(ptp
->info
));
404 int ptp_clock_unregister(struct ptp_clock
*ptp
)
406 if (ptp_vclock_in_use(ptp
)) {
407 device_for_each_child(&ptp
->dev
, NULL
, unregister_vclock
);
411 wake_up_interruptible(&ptp
->tsev_wq
);
414 kthread_cancel_delayed_work_sync(&ptp
->aux_work
);
415 kthread_destroy_worker(ptp
->kworker
);
418 /* Release the clock's resources. */
420 pps_unregister_source(ptp
->pps_source
);
422 posix_clock_unregister(&ptp
->clock
);
426 EXPORT_SYMBOL(ptp_clock_unregister
);
428 void ptp_clock_event(struct ptp_clock
*ptp
, struct ptp_clock_event
*event
)
430 struct timestamp_event_queue
*tsevq
;
431 struct pps_event_time evt
;
434 switch (event
->type
) {
436 case PTP_CLOCK_ALARM
:
439 case PTP_CLOCK_EXTTS
:
440 case PTP_CLOCK_EXTOFF
:
441 /* Enqueue timestamp on selected queues */
442 spin_lock_irqsave(&ptp
->tsevqs_lock
, flags
);
443 list_for_each_entry(tsevq
, &ptp
->tsevqs
, qlist
) {
444 if (test_bit((unsigned int)event
->index
, tsevq
->mask
))
445 enqueue_external_timestamp(tsevq
, event
);
447 spin_unlock_irqrestore(&ptp
->tsevqs_lock
, flags
);
448 wake_up_interruptible(&ptp
->tsev_wq
);
453 pps_event(ptp
->pps_source
, &evt
, PTP_PPS_EVENT
, NULL
);
456 case PTP_CLOCK_PPSUSR
:
457 pps_event(ptp
->pps_source
, &event
->pps_times
,
458 PTP_PPS_EVENT
, NULL
);
462 EXPORT_SYMBOL(ptp_clock_event
);
464 int ptp_clock_index(struct ptp_clock
*ptp
)
468 EXPORT_SYMBOL(ptp_clock_index
);
470 int ptp_find_pin(struct ptp_clock
*ptp
,
471 enum ptp_pin_function func
, unsigned int chan
)
473 struct ptp_pin_desc
*pin
= NULL
;
476 for (i
= 0; i
< ptp
->info
->n_pins
; i
++) {
477 if (ptp
->info
->pin_config
[i
].func
== func
&&
478 ptp
->info
->pin_config
[i
].chan
== chan
) {
479 pin
= &ptp
->info
->pin_config
[i
];
486 EXPORT_SYMBOL(ptp_find_pin
);
488 int ptp_find_pin_unlocked(struct ptp_clock
*ptp
,
489 enum ptp_pin_function func
, unsigned int chan
)
493 mutex_lock(&ptp
->pincfg_mux
);
495 result
= ptp_find_pin(ptp
, func
, chan
);
497 mutex_unlock(&ptp
->pincfg_mux
);
501 EXPORT_SYMBOL(ptp_find_pin_unlocked
);
503 int ptp_schedule_worker(struct ptp_clock
*ptp
, unsigned long delay
)
505 return kthread_mod_delayed_work(ptp
->kworker
, &ptp
->aux_work
, delay
);
507 EXPORT_SYMBOL(ptp_schedule_worker
);
509 void ptp_cancel_worker_sync(struct ptp_clock
*ptp
)
511 kthread_cancel_delayed_work_sync(&ptp
->aux_work
);
513 EXPORT_SYMBOL(ptp_cancel_worker_sync
);
515 /* module operations */
517 static void __exit
ptp_exit(void)
519 class_unregister(&ptp_class
);
520 unregister_chrdev_region(ptp_devt
, MINORMASK
+ 1);
521 xa_destroy(&ptp_clocks_map
);
524 static int __init
ptp_init(void)
528 err
= class_register(&ptp_class
);
530 pr_err("ptp: failed to allocate class\n");
534 err
= alloc_chrdev_region(&ptp_devt
, 0, MINORMASK
+ 1, "ptp");
536 pr_err("ptp: failed to allocate device region\n");
540 pr_info("PTP clock support registered\n");
544 class_unregister(&ptp_class
);
548 subsys_initcall(ptp_init
);
549 module_exit(ptp_exit
);
551 MODULE_AUTHOR("Richard Cochran <richardcochran@gmail.com>");
552 MODULE_DESCRIPTION("PTP clocks support");
553 MODULE_LICENSE("GPL");