4 * Core support: hpsb_packet management, packet handling and forwarding to
5 * highlevel or lowlevel code
7 * Copyright (C) 1999, 2000 Andreas E. Bombe
8 * 2002 Manfred Weihs <weihs@ict.tuwien.ac.at>
10 * This code is licensed under the GPL. See the file COPYING in the root
11 * directory of the kernel sources for details.
16 * Manfred Weihs <weihs@ict.tuwien.ac.at>
17 * loopback functionality in hpsb_send_packet
18 * allow highlevel drivers to disable automatic response generation
19 * and to generate responses themselves (deferred)
23 #include <linux/config.h>
24 #include <linux/kernel.h>
25 #include <linux/list.h>
26 #include <linux/string.h>
27 #include <linux/init.h>
28 #include <linux/slab.h>
29 #include <linux/interrupt.h>
30 #include <linux/module.h>
31 #include <linux/moduleparam.h>
32 #include <linux/bitops.h>
33 #include <linux/kdev_t.h>
34 #include <linux/skbuff.h>
35 #include <linux/suspend.h>
37 #include <asm/byteorder.h>
38 #include <asm/semaphore.h>
40 #include "ieee1394_types.h"
43 #include "ieee1394_core.h"
44 #include "highlevel.h"
45 #include "ieee1394_transactions.h"
50 #include "config_roms.h"
53 * Disable the nodemgr detection and config rom reading functionality.
55 static int disable_nodemgr;
56 module_param(disable_nodemgr, int, 0444);
57 MODULE_PARM_DESC(disable_nodemgr, "Disable nodemgr functionality.");
59 /* Disable Isochronous Resource Manager functionality */
60 int hpsb_disable_irm = 0;
61 module_param_named(disable_irm, hpsb_disable_irm, bool, 0);
62 MODULE_PARM_DESC(disable_irm,
63 "Disable Isochronous Resource Manager functionality.");
65 /* We are GPL, so treat us special */
66 MODULE_LICENSE("GPL");
68 /* Some globals used */
69 const char *hpsb_speedto_str[] = { "S100", "S200", "S400", "S800", "S1600", "S3200" };
70 struct class *hpsb_protocol_class;
72 #ifdef CONFIG_IEEE1394_VERBOSEDEBUG
73 static void dump_packet(const char *text, quadlet_t *data, int size, int speed)
78 size = (size > 4 ? 4 : size);
80 printk(KERN_DEBUG "ieee1394: %s", text);
81 if (speed > -1 && speed < 6)
82 printk(" at %s", hpsb_speedto_str[speed]);
84 for (i = 0; i < size; i++)
85 printk(" %08x", data[i]);
89 #define dump_packet(a,b,c,d)
92 static void abort_requests(struct hpsb_host *host);
93 static void queue_packet_complete(struct hpsb_packet *packet);
97 * hpsb_set_packet_complete_task - set the task that runs when a packet
98 * completes. You cannot call this more than once on a single packet
101 * @packet: the packet whose completion we want the task added to
102 * @routine: function to call
103 * @data: data (if any) to pass to the above function
105 void hpsb_set_packet_complete_task(struct hpsb_packet *packet,
106 void (*routine)(void *), void *data)
108 WARN_ON(packet->complete_routine != NULL);
109 packet->complete_routine = routine;
110 packet->complete_data = data;
115 * hpsb_alloc_packet - allocate new packet structure
116 * @data_size: size of the data block to be allocated
118 * This function allocates, initializes and returns a new &struct hpsb_packet.
119 * It can be used in interrupt context. A header block is always included, its
120 * size is big enough to contain all possible 1394 headers. The data block is
121 * only allocated when @data_size is not zero.
123 * For packets for which responses will be received the @data_size has to be big
124 * enough to contain the response's data block since no further allocation
125 * occurs at response matching time.
127 * The packet's generation value will be set to the current generation number
128 * for ease of use. Remember to overwrite it with your own recorded generation
129 * number if you can not be sure that your code will not race with a bus reset.
131 * Return value: A pointer to a &struct hpsb_packet or NULL on allocation
134 struct hpsb_packet *hpsb_alloc_packet(size_t data_size)
136 struct hpsb_packet *packet = NULL;
139 data_size = ((data_size + 3) & ~3);
141 skb = alloc_skb(data_size + sizeof(*packet), GFP_ATOMIC);
145 memset(skb->data, 0, data_size + sizeof(*packet));
147 packet = (struct hpsb_packet *)skb->data;
150 packet->header = packet->embedded_header;
151 packet->state = hpsb_unused;
152 packet->generation = -1;
153 INIT_LIST_HEAD(&packet->driver_list);
154 atomic_set(&packet->refcnt, 1);
157 packet->data = (quadlet_t *)(skb->data + sizeof(*packet));
158 packet->data_size = data_size;
166 * hpsb_free_packet - free packet and data associated with it
167 * @packet: packet to free (is NULL safe)
169 * This function will free packet->data and finally the packet itself.
171 void hpsb_free_packet(struct hpsb_packet *packet)
173 if (packet && atomic_dec_and_test(&packet->refcnt)) {
174 BUG_ON(!list_empty(&packet->driver_list));
175 kfree_skb(packet->skb);
180 int hpsb_reset_bus(struct hpsb_host *host, int type)
182 if (!host->in_bus_reset) {
183 host->driver->devctl(host, RESET_BUS, type);
191 int hpsb_bus_reset(struct hpsb_host *host)
193 if (host->in_bus_reset) {
194 HPSB_NOTICE("%s called while bus reset already in progress",
199 abort_requests(host);
200 host->in_bus_reset = 1;
203 host->busmgr_id = -1;
206 host->node_count = 0;
207 host->selfid_count = 0;
214 * Verify num_of_selfids SelfIDs and return number of nodes. Return zero in
215 * case verification failed.
217 static int check_selfids(struct hpsb_host *host)
220 int rest_of_selfids = host->selfid_count;
221 struct selfid *sid = (struct selfid *)host->topology_map;
222 struct ext_selfid *esid;
225 host->nodes_active = 0;
227 while (rest_of_selfids--) {
228 if (!sid->extended) {
232 if (sid->phy_id != nodeid) {
233 HPSB_INFO("SelfIDs failed monotony check with "
238 if (sid->link_active) {
239 host->nodes_active++;
241 host->irm_id = LOCAL_BUS | sid->phy_id;
244 esid = (struct ext_selfid *)sid;
246 if ((esid->phy_id != nodeid)
247 || (esid->seq_nr != esid_seq)) {
248 HPSB_INFO("SelfIDs failed monotony check with "
249 "%d/%d", esid->phy_id, esid->seq_nr);
257 esid = (struct ext_selfid *)(sid - 1);
258 while (esid->extended) {
259 if ((esid->porta == SELFID_PORT_PARENT) ||
260 (esid->portb == SELFID_PORT_PARENT) ||
261 (esid->portc == SELFID_PORT_PARENT) ||
262 (esid->portd == SELFID_PORT_PARENT) ||
263 (esid->porte == SELFID_PORT_PARENT) ||
264 (esid->portf == SELFID_PORT_PARENT) ||
265 (esid->portg == SELFID_PORT_PARENT) ||
266 (esid->porth == SELFID_PORT_PARENT)) {
267 HPSB_INFO("SelfIDs failed root check on "
274 sid = (struct selfid *)esid;
275 if ((sid->port0 == SELFID_PORT_PARENT) ||
276 (sid->port1 == SELFID_PORT_PARENT) ||
277 (sid->port2 == SELFID_PORT_PARENT)) {
278 HPSB_INFO("SelfIDs failed root check");
282 host->node_count = nodeid + 1;
286 static void build_speed_map(struct hpsb_host *host, int nodecount)
288 u8 speedcap[nodecount];
289 u8 cldcnt[nodecount];
290 u8 *map = host->speed_map;
292 struct ext_selfid *esid;
295 for (i = 0; i < (nodecount * 64); i += 64) {
296 for (j = 0; j < nodecount; j++) {
297 map[i+j] = IEEE1394_SPEED_MAX;
301 for (i = 0; i < nodecount; i++) {
305 /* find direct children count and speed */
306 for (sid = (struct selfid *)&host->topology_map[host->selfid_count-1],
308 (void *)sid >= (void *)host->topology_map; sid--) {
310 esid = (struct ext_selfid *)sid;
312 if (esid->porta == SELFID_PORT_CHILD) cldcnt[n]++;
313 if (esid->portb == SELFID_PORT_CHILD) cldcnt[n]++;
314 if (esid->portc == SELFID_PORT_CHILD) cldcnt[n]++;
315 if (esid->portd == SELFID_PORT_CHILD) cldcnt[n]++;
316 if (esid->porte == SELFID_PORT_CHILD) cldcnt[n]++;
317 if (esid->portf == SELFID_PORT_CHILD) cldcnt[n]++;
318 if (esid->portg == SELFID_PORT_CHILD) cldcnt[n]++;
319 if (esid->porth == SELFID_PORT_CHILD) cldcnt[n]++;
321 if (sid->port0 == SELFID_PORT_CHILD) cldcnt[n]++;
322 if (sid->port1 == SELFID_PORT_CHILD) cldcnt[n]++;
323 if (sid->port2 == SELFID_PORT_CHILD) cldcnt[n]++;
325 speedcap[n] = sid->speed;
330 /* set self mapping */
331 for (i = 0; i < nodecount; i++) {
332 map[64*i + i] = speedcap[i];
335 /* fix up direct children count to total children count;
336 * also fix up speedcaps for sibling and parent communication */
337 for (i = 1; i < nodecount; i++) {
338 for (j = cldcnt[i], n = i - 1; j > 0; j--) {
339 cldcnt[i] += cldcnt[n];
340 speedcap[n] = min(speedcap[n], speedcap[i]);
345 for (n = 0; n < nodecount; n++) {
346 for (i = n - cldcnt[n]; i <= n; i++) {
347 for (j = 0; j < (n - cldcnt[n]); j++) {
348 map[j*64 + i] = map[i*64 + j] =
349 min(map[i*64 + j], speedcap[n]);
351 for (j = n + 1; j < nodecount; j++) {
352 map[j*64 + i] = map[i*64 + j] =
353 min(map[i*64 + j], speedcap[n]);
360 void hpsb_selfid_received(struct hpsb_host *host, quadlet_t sid)
362 if (host->in_bus_reset) {
363 HPSB_VERBOSE("Including SelfID 0x%x", sid);
364 host->topology_map[host->selfid_count++] = sid;
366 HPSB_NOTICE("Spurious SelfID packet (0x%08x) received from bus %d",
367 sid, NODEID_TO_BUS(host->node_id));
371 void hpsb_selfid_complete(struct hpsb_host *host, int phyid, int isroot)
373 if (!host->in_bus_reset)
374 HPSB_NOTICE("SelfID completion called outside of bus reset!");
376 host->node_id = LOCAL_BUS | phyid;
377 host->is_root = isroot;
379 if (!check_selfids(host)) {
380 if (host->reset_retries++ < 20) {
381 /* selfid stage did not complete without error */
382 HPSB_NOTICE("Error in SelfID stage, resetting");
383 host->in_bus_reset = 0;
384 /* this should work from ohci1394 now... */
385 hpsb_reset_bus(host, LONG_RESET);
388 HPSB_NOTICE("Stopping out-of-control reset loop");
389 HPSB_NOTICE("Warning - topology map and speed map will not be valid");
390 host->reset_retries = 0;
393 host->reset_retries = 0;
394 build_speed_map(host, host->node_count);
397 HPSB_VERBOSE("selfid_complete called with successful SelfID stage "
398 "... irm_id: 0x%X node_id: 0x%X",host->irm_id,host->node_id);
400 /* irm_id is kept up to date by check_selfids() */
401 if (host->irm_id == host->node_id) {
409 host->driver->devctl(host, ACT_CYCLE_MASTER, 1);
412 atomic_inc(&host->generation);
413 host->in_bus_reset = 0;
414 highlevel_host_reset(host);
418 void hpsb_packet_sent(struct hpsb_host *host, struct hpsb_packet *packet,
423 spin_lock_irqsave(&host->pending_packet_queue.lock, flags);
425 packet->ack_code = ackcode;
427 if (packet->no_waiter || packet->state == hpsb_complete) {
428 /* if packet->no_waiter, must not have a tlabel allocated */
429 spin_unlock_irqrestore(&host->pending_packet_queue.lock, flags);
430 hpsb_free_packet(packet);
434 atomic_dec(&packet->refcnt); /* drop HC's reference */
435 /* here the packet must be on the host->pending_packet_queue */
437 if (ackcode != ACK_PENDING || !packet->expect_response) {
438 packet->state = hpsb_complete;
439 __skb_unlink(packet->skb, &host->pending_packet_queue);
440 spin_unlock_irqrestore(&host->pending_packet_queue.lock, flags);
441 queue_packet_complete(packet);
445 packet->state = hpsb_pending;
446 packet->sendtime = jiffies;
448 spin_unlock_irqrestore(&host->pending_packet_queue.lock, flags);
450 mod_timer(&host->timeout, jiffies + host->timeout_interval);
454 * hpsb_send_phy_config - transmit a PHY configuration packet on the bus
455 * @host: host that PHY config packet gets sent through
456 * @rootid: root whose force_root bit should get set (-1 = don't set force_root)
457 * @gapcnt: gap count value to set (-1 = don't set gap count)
459 * This function sends a PHY config packet on the bus through the specified host.
461 * Return value: 0 for success or error number otherwise.
463 int hpsb_send_phy_config(struct hpsb_host *host, int rootid, int gapcnt)
465 struct hpsb_packet *packet;
468 if (rootid >= ALL_NODES || rootid < -1 || gapcnt > 0x3f || gapcnt < -1 ||
469 (rootid == -1 && gapcnt == -1)) {
470 HPSB_DEBUG("Invalid Parameter: rootid = %d gapcnt = %d",
475 packet = hpsb_alloc_packet(0);
480 packet->header_size = 8;
481 packet->data_size = 0;
482 packet->expect_response = 0;
483 packet->no_waiter = 0;
484 packet->type = hpsb_raw;
485 packet->header[0] = 0;
487 packet->header[0] |= rootid << 24 | 1 << 23;
489 packet->header[0] |= gapcnt << 16 | 1 << 22;
491 packet->header[1] = ~packet->header[0];
493 packet->generation = get_hpsb_generation(host);
495 retval = hpsb_send_packet_and_wait(packet);
496 hpsb_free_packet(packet);
502 * hpsb_send_packet - transmit a packet on the bus
503 * @packet: packet to send
505 * The packet is sent through the host specified in the packet->host field.
506 * Before sending, the packet's transmit speed is automatically determined
507 * using the local speed map when it is an async, non-broadcast packet.
509 * Possibilities for failure are that host is either not initialized, in bus
510 * reset, the packet's generation number doesn't match the current generation
511 * number or the host reports a transmit error.
513 * Return value: 0 on success, negative errno on failure.
515 int hpsb_send_packet(struct hpsb_packet *packet)
517 struct hpsb_host *host = packet->host;
519 if (host->is_shutdown)
521 if (host->in_bus_reset ||
522 (packet->generation != get_hpsb_generation(host)))
525 packet->state = hpsb_queued;
527 /* This just seems silly to me */
528 WARN_ON(packet->no_waiter && packet->expect_response);
530 if (!packet->no_waiter || packet->expect_response) {
531 atomic_inc(&packet->refcnt);
532 /* Set the initial "sendtime" to 10 seconds from now, to
533 prevent premature expiry. If a packet takes more than
534 10 seconds to hit the wire, we have bigger problems :) */
535 packet->sendtime = jiffies + 10 * HZ;
536 skb_queue_tail(&host->pending_packet_queue, packet->skb);
539 if (packet->node_id == host->node_id) {
540 /* it is a local request, so handle it locally */
543 size_t size = packet->data_size + packet->header_size;
545 data = kmalloc(size, GFP_ATOMIC);
547 HPSB_ERR("unable to allocate memory for concatenating header and data");
551 memcpy(data, packet->header, packet->header_size);
553 if (packet->data_size)
554 memcpy(((u8*)data) + packet->header_size, packet->data, packet->data_size);
556 dump_packet("send packet local", packet->header, packet->header_size, -1);
558 hpsb_packet_sent(host, packet, packet->expect_response ? ACK_PENDING : ACK_COMPLETE);
559 hpsb_packet_received(host, data, size, 0);
566 if (packet->type == hpsb_async && packet->node_id != ALL_NODES) {
568 host->speed_map[NODEID_TO_NODE(host->node_id) * 64
569 + NODEID_TO_NODE(packet->node_id)];
572 dump_packet("send packet", packet->header, packet->header_size, packet->speed_code);
574 return host->driver->transmit_packet(host, packet);
577 /* We could just use complete() directly as the packet complete
578 * callback, but this is more typesafe, in the sense that we get a
579 * compiler error if the prototype for complete() changes. */
581 static void complete_packet(void *data)
583 complete((struct completion *) data);
586 int hpsb_send_packet_and_wait(struct hpsb_packet *packet)
588 struct completion done;
591 init_completion(&done);
592 hpsb_set_packet_complete_task(packet, complete_packet, &done);
593 retval = hpsb_send_packet(packet);
595 wait_for_completion(&done);
600 static void send_packet_nocare(struct hpsb_packet *packet)
602 if (hpsb_send_packet(packet) < 0) {
603 hpsb_free_packet(packet);
608 static void handle_packet_response(struct hpsb_host *host, int tcode,
609 quadlet_t *data, size_t size)
611 struct hpsb_packet *packet = NULL;
617 tlabel = (data[0] >> 10) & 0x3f;
619 spin_lock_irqsave(&host->pending_packet_queue.lock, flags);
621 skb_queue_walk(&host->pending_packet_queue, skb) {
622 packet = (struct hpsb_packet *)skb->data;
623 if ((packet->tlabel == tlabel)
624 && (packet->node_id == (data[1] >> 16))){
631 if (packet == NULL) {
632 HPSB_DEBUG("unsolicited response packet received - no tlabel match");
633 dump_packet("contents", data, 16, -1);
634 spin_unlock_irqrestore(&host->pending_packet_queue.lock, flags);
638 switch (packet->tcode) {
641 if (tcode != TCODE_WRITE_RESPONSE)
644 memcpy(packet->header, data, 12);
647 if (tcode != TCODE_READQ_RESPONSE)
650 memcpy(packet->header, data, 16);
653 if (tcode != TCODE_READB_RESPONSE)
656 BUG_ON(packet->skb->len - sizeof(*packet) < size - 16);
657 memcpy(packet->header, data, 16);
658 memcpy(packet->data, data + 4, size - 16);
660 case TCODE_LOCK_REQUEST:
661 if (tcode != TCODE_LOCK_RESPONSE)
664 size = min((size - 16), (size_t)8);
665 BUG_ON(packet->skb->len - sizeof(*packet) < size);
666 memcpy(packet->header, data, 16);
667 memcpy(packet->data, data + 4, size);
672 spin_unlock_irqrestore(&host->pending_packet_queue.lock, flags);
673 HPSB_INFO("unsolicited response packet received - tcode mismatch");
674 dump_packet("contents", data, 16, -1);
678 __skb_unlink(skb, &host->pending_packet_queue);
680 if (packet->state == hpsb_queued) {
681 packet->sendtime = jiffies;
682 packet->ack_code = ACK_PENDING;
685 packet->state = hpsb_complete;
686 spin_unlock_irqrestore(&host->pending_packet_queue.lock, flags);
688 queue_packet_complete(packet);
692 static struct hpsb_packet *create_reply_packet(struct hpsb_host *host,
693 quadlet_t *data, size_t dsize)
695 struct hpsb_packet *p;
697 p = hpsb_alloc_packet(dsize);
698 if (unlikely(p == NULL)) {
699 /* FIXME - send data_error response */
703 p->type = hpsb_async;
704 p->state = hpsb_unused;
706 p->node_id = data[1] >> 16;
707 p->tlabel = (data[0] >> 10) & 0x3f;
710 p->generation = get_hpsb_generation(host);
713 p->data[dsize / 4] = 0;
718 #define PREP_ASYNC_HEAD_RCODE(tc) \
719 packet->tcode = tc; \
720 packet->header[0] = (packet->node_id << 16) | (packet->tlabel << 10) \
721 | (1 << 8) | (tc << 4); \
722 packet->header[1] = (packet->host->node_id << 16) | (rcode << 12); \
723 packet->header[2] = 0
725 static void fill_async_readquad_resp(struct hpsb_packet *packet, int rcode,
728 PREP_ASYNC_HEAD_RCODE(TCODE_READQ_RESPONSE);
729 packet->header[3] = data;
730 packet->header_size = 16;
731 packet->data_size = 0;
734 static void fill_async_readblock_resp(struct hpsb_packet *packet, int rcode,
737 if (rcode != RCODE_COMPLETE)
740 PREP_ASYNC_HEAD_RCODE(TCODE_READB_RESPONSE);
741 packet->header[3] = length << 16;
742 packet->header_size = 16;
743 packet->data_size = length + (length % 4 ? 4 - (length % 4) : 0);
746 static void fill_async_write_resp(struct hpsb_packet *packet, int rcode)
748 PREP_ASYNC_HEAD_RCODE(TCODE_WRITE_RESPONSE);
749 packet->header[2] = 0;
750 packet->header_size = 12;
751 packet->data_size = 0;
754 static void fill_async_lock_resp(struct hpsb_packet *packet, int rcode, int extcode,
757 if (rcode != RCODE_COMPLETE)
760 PREP_ASYNC_HEAD_RCODE(TCODE_LOCK_RESPONSE);
761 packet->header[3] = (length << 16) | extcode;
762 packet->header_size = 16;
763 packet->data_size = length;
766 #define PREP_REPLY_PACKET(length) \
767 packet = create_reply_packet(host, data, length); \
768 if (packet == NULL) break
770 static void handle_incoming_packet(struct hpsb_host *host, int tcode,
771 quadlet_t *data, size_t size, int write_acked)
773 struct hpsb_packet *packet;
774 int length, rcode, extcode;
776 nodeid_t source = data[1] >> 16;
777 nodeid_t dest = data[0] >> 16;
778 u16 flags = (u16) data[0];
781 /* big FIXME - no error checking is done for an out of bounds length */
785 addr = (((u64)(data[1] & 0xffff)) << 32) | data[2];
786 rcode = highlevel_write(host, source, dest, data+3,
790 && (NODEID_TO_NODE(data[0] >> 16) != NODE_MASK)
792 /* not a broadcast write, reply */
793 PREP_REPLY_PACKET(0);
794 fill_async_write_resp(packet, rcode);
795 send_packet_nocare(packet);
800 addr = (((u64)(data[1] & 0xffff)) << 32) | data[2];
801 rcode = highlevel_write(host, source, dest, data+4,
802 addr, data[3]>>16, flags);
805 && (NODEID_TO_NODE(data[0] >> 16) != NODE_MASK)
807 /* not a broadcast write, reply */
808 PREP_REPLY_PACKET(0);
809 fill_async_write_resp(packet, rcode);
810 send_packet_nocare(packet);
815 addr = (((u64)(data[1] & 0xffff)) << 32) | data[2];
816 rcode = highlevel_read(host, source, &buffer, addr, 4, flags);
819 PREP_REPLY_PACKET(0);
820 fill_async_readquad_resp(packet, rcode, buffer);
821 send_packet_nocare(packet);
826 length = data[3] >> 16;
827 PREP_REPLY_PACKET(length);
829 addr = (((u64)(data[1] & 0xffff)) << 32) | data[2];
830 rcode = highlevel_read(host, source, packet->data, addr,
834 fill_async_readblock_resp(packet, rcode, length);
835 send_packet_nocare(packet);
837 hpsb_free_packet(packet);
841 case TCODE_LOCK_REQUEST:
842 length = data[3] >> 16;
843 extcode = data[3] & 0xffff;
844 addr = (((u64)(data[1] & 0xffff)) << 32) | data[2];
846 PREP_REPLY_PACKET(8);
848 if ((extcode == 0) || (extcode >= 7)) {
849 /* let switch default handle error */
855 rcode = highlevel_lock(host, source, packet->data, addr,
856 data[4], 0, extcode,flags);
857 fill_async_lock_resp(packet, rcode, extcode, 4);
860 if ((extcode != EXTCODE_FETCH_ADD)
861 && (extcode != EXTCODE_LITTLE_ADD)) {
862 rcode = highlevel_lock(host, source,
866 fill_async_lock_resp(packet, rcode, extcode, 4);
868 rcode = highlevel_lock64(host, source,
869 (octlet_t *)packet->data, addr,
870 *(octlet_t *)(data + 4), 0ULL,
872 fill_async_lock_resp(packet, rcode, extcode, 8);
876 rcode = highlevel_lock64(host, source,
877 (octlet_t *)packet->data, addr,
878 *(octlet_t *)(data + 6),
879 *(octlet_t *)(data + 4),
881 fill_async_lock_resp(packet, rcode, extcode, 8);
884 rcode = RCODE_TYPE_ERROR;
885 fill_async_lock_resp(packet, rcode,
890 send_packet_nocare(packet);
892 hpsb_free_packet(packet);
898 #undef PREP_REPLY_PACKET
901 void hpsb_packet_received(struct hpsb_host *host, quadlet_t *data, size_t size,
906 if (host->in_bus_reset) {
907 HPSB_INFO("received packet during reset; ignoring");
911 dump_packet("received packet", data, size, -1);
913 tcode = (data[0] >> 4) & 0xf;
916 case TCODE_WRITE_RESPONSE:
917 case TCODE_READQ_RESPONSE:
918 case TCODE_READB_RESPONSE:
919 case TCODE_LOCK_RESPONSE:
920 handle_packet_response(host, tcode, data, size);
927 case TCODE_LOCK_REQUEST:
928 handle_incoming_packet(host, tcode, data, size, write_acked);
933 highlevel_iso_receive(host, data, size);
936 case TCODE_CYCLE_START:
937 /* simply ignore this packet if it is passed on */
941 HPSB_NOTICE("received packet with bogus transaction code %d",
948 static void abort_requests(struct hpsb_host *host)
950 struct hpsb_packet *packet;
953 host->driver->devctl(host, CANCEL_REQUESTS, 0);
955 while ((skb = skb_dequeue(&host->pending_packet_queue)) != NULL) {
956 packet = (struct hpsb_packet *)skb->data;
958 packet->state = hpsb_complete;
959 packet->ack_code = ACKX_ABORTED;
960 queue_packet_complete(packet);
964 void abort_timedouts(unsigned long __opaque)
966 struct hpsb_host *host = (struct hpsb_host *)__opaque;
968 struct hpsb_packet *packet;
970 unsigned long expire;
972 spin_lock_irqsave(&host->csr.lock, flags);
973 expire = host->csr.expire;
974 spin_unlock_irqrestore(&host->csr.lock, flags);
976 /* Hold the lock around this, since we aren't dequeuing all
977 * packets, just ones we need. */
978 spin_lock_irqsave(&host->pending_packet_queue.lock, flags);
980 while (!skb_queue_empty(&host->pending_packet_queue)) {
981 skb = skb_peek(&host->pending_packet_queue);
983 packet = (struct hpsb_packet *)skb->data;
985 if (time_before(packet->sendtime + expire, jiffies)) {
986 __skb_unlink(skb, &host->pending_packet_queue);
987 packet->state = hpsb_complete;
988 packet->ack_code = ACKX_TIMEOUT;
989 queue_packet_complete(packet);
991 /* Since packets are added to the tail, the oldest
992 * ones are first, always. When we get to one that
993 * isn't timed out, the rest aren't either. */
998 if (!skb_queue_empty(&host->pending_packet_queue))
999 mod_timer(&host->timeout, jiffies + host->timeout_interval);
1001 spin_unlock_irqrestore(&host->pending_packet_queue.lock, flags);
1005 /* Kernel thread and vars, which handles packets that are completed. Only
1006 * packets that have a "complete" function are sent here. This way, the
1007 * completion is run out of kernel context, and doesn't block the rest of
1009 static int khpsbpkt_pid = -1, khpsbpkt_kill;
1010 static DECLARE_COMPLETION(khpsbpkt_complete);
1011 static struct sk_buff_head hpsbpkt_queue;
1012 static DECLARE_MUTEX_LOCKED(khpsbpkt_sig);
1015 static void queue_packet_complete(struct hpsb_packet *packet)
1017 if (packet->no_waiter) {
1018 hpsb_free_packet(packet);
1021 if (packet->complete_routine != NULL) {
1022 skb_queue_tail(&hpsbpkt_queue, packet->skb);
1024 /* Signal the kernel thread to handle this */
1030 static int hpsbpkt_thread(void *__hi)
1032 struct sk_buff *skb;
1033 struct hpsb_packet *packet;
1034 void (*complete_routine)(void*);
1035 void *complete_data;
1037 daemonize("khpsbpkt");
1040 if (down_interruptible(&khpsbpkt_sig)) {
1041 if (try_to_freeze())
1043 printk("khpsbpkt: received unexpected signal?!\n" );
1050 while ((skb = skb_dequeue(&hpsbpkt_queue)) != NULL) {
1051 packet = (struct hpsb_packet *)skb->data;
1053 complete_routine = packet->complete_routine;
1054 complete_data = packet->complete_data;
1056 packet->complete_routine = packet->complete_data = NULL;
1058 complete_routine(complete_data);
1062 complete_and_exit(&khpsbpkt_complete, 0);
1065 static int __init ieee1394_init(void)
1069 skb_queue_head_init(&hpsbpkt_queue);
1071 /* non-fatal error */
1072 if (hpsb_init_config_roms()) {
1073 HPSB_ERR("Failed to initialize some config rom entries.\n");
1074 HPSB_ERR("Some features may not be available\n");
1077 khpsbpkt_pid = kernel_thread(hpsbpkt_thread, NULL, CLONE_KERNEL);
1078 if (khpsbpkt_pid < 0) {
1079 HPSB_ERR("Failed to start hpsbpkt thread!\n");
1081 goto exit_cleanup_config_roms;
1084 if (register_chrdev_region(IEEE1394_CORE_DEV, 256, "ieee1394")) {
1085 HPSB_ERR("unable to register character device major %d!\n", IEEE1394_MAJOR);
1087 goto exit_release_kernel_thread;
1090 /* actually this is a non-fatal error */
1091 ret = devfs_mk_dir("ieee1394");
1093 HPSB_ERR("unable to make devfs dir for device major %d!\n", IEEE1394_MAJOR);
1094 goto release_chrdev;
1097 ret = bus_register(&ieee1394_bus_type);
1099 HPSB_INFO("bus register failed");
1103 for (i = 0; fw_bus_attrs[i]; i++) {
1104 ret = bus_create_file(&ieee1394_bus_type, fw_bus_attrs[i]);
1107 bus_remove_file(&ieee1394_bus_type,
1110 bus_unregister(&ieee1394_bus_type);
1115 ret = class_register(&hpsb_host_class);
1117 goto release_all_bus;
1119 hpsb_protocol_class = class_create(THIS_MODULE, "ieee1394_protocol");
1120 if (IS_ERR(hpsb_protocol_class)) {
1121 ret = PTR_ERR(hpsb_protocol_class);
1122 goto release_class_host;
1127 HPSB_INFO("init csr failed");
1129 goto release_class_protocol;
1132 if (disable_nodemgr) {
1133 HPSB_INFO("nodemgr and IRM functionality disabled");
1134 /* We shouldn't contend for IRM with nodemgr disabled, since
1135 nodemgr implements functionality required of ieee1394a-2000
1137 hpsb_disable_irm = 1;
1142 if (hpsb_disable_irm) {
1143 HPSB_INFO("IRM functionality disabled");
1146 ret = init_ieee1394_nodemgr();
1148 HPSB_INFO("init nodemgr failed");
1156 release_class_protocol:
1157 class_destroy(hpsb_protocol_class);
1159 class_unregister(&hpsb_host_class);
1161 for (i = 0; fw_bus_attrs[i]; i++)
1162 bus_remove_file(&ieee1394_bus_type, fw_bus_attrs[i]);
1163 bus_unregister(&ieee1394_bus_type);
1165 devfs_remove("ieee1394");
1167 unregister_chrdev_region(IEEE1394_CORE_DEV, 256);
1168 exit_release_kernel_thread:
1169 if (khpsbpkt_pid >= 0) {
1170 kill_proc(khpsbpkt_pid, SIGTERM, 1);
1171 wait_for_completion(&khpsbpkt_complete);
1173 exit_cleanup_config_roms:
1174 hpsb_cleanup_config_roms();
1178 static void __exit ieee1394_cleanup(void)
1182 if (!disable_nodemgr)
1183 cleanup_ieee1394_nodemgr();
1187 class_destroy(hpsb_protocol_class);
1188 class_unregister(&hpsb_host_class);
1189 for (i = 0; fw_bus_attrs[i]; i++)
1190 bus_remove_file(&ieee1394_bus_type, fw_bus_attrs[i]);
1191 bus_unregister(&ieee1394_bus_type);
1193 if (khpsbpkt_pid >= 0) {
1197 wait_for_completion(&khpsbpkt_complete);
1200 hpsb_cleanup_config_roms();
1202 unregister_chrdev_region(IEEE1394_CORE_DEV, 256);
1203 devfs_remove("ieee1394");
1206 module_init(ieee1394_init);
1207 module_exit(ieee1394_cleanup);
1209 /* Exported symbols */
1212 EXPORT_SYMBOL(hpsb_alloc_host);
1213 EXPORT_SYMBOL(hpsb_add_host);
1214 EXPORT_SYMBOL(hpsb_remove_host);
1215 EXPORT_SYMBOL(hpsb_update_config_rom_image);
1217 /** ieee1394_core.c **/
1218 EXPORT_SYMBOL(hpsb_speedto_str);
1219 EXPORT_SYMBOL(hpsb_protocol_class);
1220 EXPORT_SYMBOL(hpsb_set_packet_complete_task);
1221 EXPORT_SYMBOL(hpsb_alloc_packet);
1222 EXPORT_SYMBOL(hpsb_free_packet);
1223 EXPORT_SYMBOL(hpsb_send_packet);
1224 EXPORT_SYMBOL(hpsb_reset_bus);
1225 EXPORT_SYMBOL(hpsb_bus_reset);
1226 EXPORT_SYMBOL(hpsb_selfid_received);
1227 EXPORT_SYMBOL(hpsb_selfid_complete);
1228 EXPORT_SYMBOL(hpsb_packet_sent);
1229 EXPORT_SYMBOL(hpsb_packet_received);
1230 EXPORT_SYMBOL_GPL(hpsb_disable_irm);
1231 #ifdef CONFIG_IEEE1394_EXPORT_FULL_API
1232 EXPORT_SYMBOL(hpsb_send_phy_config);
1233 EXPORT_SYMBOL(hpsb_send_packet_and_wait);
1236 /** ieee1394_transactions.c **/
1237 EXPORT_SYMBOL(hpsb_get_tlabel);
1238 EXPORT_SYMBOL(hpsb_free_tlabel);
1239 EXPORT_SYMBOL(hpsb_make_readpacket);
1240 EXPORT_SYMBOL(hpsb_make_writepacket);
1241 EXPORT_SYMBOL(hpsb_make_streampacket);
1242 EXPORT_SYMBOL(hpsb_make_lockpacket);
1243 EXPORT_SYMBOL(hpsb_make_lock64packet);
1244 EXPORT_SYMBOL(hpsb_make_phypacket);
1245 EXPORT_SYMBOL(hpsb_make_isopacket);
1246 EXPORT_SYMBOL(hpsb_read);
1247 EXPORT_SYMBOL(hpsb_write);
1248 EXPORT_SYMBOL(hpsb_packet_success);
1251 EXPORT_SYMBOL(hpsb_register_highlevel);
1252 EXPORT_SYMBOL(hpsb_unregister_highlevel);
1253 EXPORT_SYMBOL(hpsb_register_addrspace);
1254 EXPORT_SYMBOL(hpsb_unregister_addrspace);
1255 EXPORT_SYMBOL(hpsb_allocate_and_register_addrspace);
1256 EXPORT_SYMBOL(hpsb_listen_channel);
1257 EXPORT_SYMBOL(hpsb_unlisten_channel);
1258 EXPORT_SYMBOL(hpsb_get_hostinfo);
1259 EXPORT_SYMBOL(hpsb_create_hostinfo);
1260 EXPORT_SYMBOL(hpsb_destroy_hostinfo);
1261 EXPORT_SYMBOL(hpsb_set_hostinfo_key);
1262 EXPORT_SYMBOL(hpsb_get_hostinfo_bykey);
1263 EXPORT_SYMBOL(hpsb_set_hostinfo);
1264 EXPORT_SYMBOL(highlevel_host_reset);
1265 #ifdef CONFIG_IEEE1394_EXPORT_FULL_API
1266 EXPORT_SYMBOL(highlevel_add_host);
1267 EXPORT_SYMBOL(highlevel_remove_host);
1271 EXPORT_SYMBOL(hpsb_node_fill_packet);
1272 EXPORT_SYMBOL(hpsb_node_write);
1273 EXPORT_SYMBOL(hpsb_register_protocol);
1274 EXPORT_SYMBOL(hpsb_unregister_protocol);
1275 EXPORT_SYMBOL(ieee1394_bus_type);
1276 #ifdef CONFIG_IEEE1394_EXPORT_FULL_API
1277 EXPORT_SYMBOL(nodemgr_for_each_host);
1281 EXPORT_SYMBOL(hpsb_update_config_rom);
1284 EXPORT_SYMBOL(dma_prog_region_init);
1285 EXPORT_SYMBOL(dma_prog_region_alloc);
1286 EXPORT_SYMBOL(dma_prog_region_free);
1287 EXPORT_SYMBOL(dma_region_init);
1288 EXPORT_SYMBOL(dma_region_alloc);
1289 EXPORT_SYMBOL(dma_region_free);
1290 EXPORT_SYMBOL(dma_region_sync_for_cpu);
1291 EXPORT_SYMBOL(dma_region_sync_for_device);
1292 EXPORT_SYMBOL(dma_region_mmap);
1293 EXPORT_SYMBOL(dma_region_offset_to_bus);
1296 EXPORT_SYMBOL(hpsb_iso_xmit_init);
1297 EXPORT_SYMBOL(hpsb_iso_recv_init);
1298 EXPORT_SYMBOL(hpsb_iso_xmit_start);
1299 EXPORT_SYMBOL(hpsb_iso_recv_start);
1300 EXPORT_SYMBOL(hpsb_iso_recv_listen_channel);
1301 EXPORT_SYMBOL(hpsb_iso_recv_unlisten_channel);
1302 EXPORT_SYMBOL(hpsb_iso_recv_set_channel_mask);
1303 EXPORT_SYMBOL(hpsb_iso_stop);
1304 EXPORT_SYMBOL(hpsb_iso_shutdown);
1305 EXPORT_SYMBOL(hpsb_iso_xmit_queue_packet);
1306 EXPORT_SYMBOL(hpsb_iso_xmit_sync);
1307 EXPORT_SYMBOL(hpsb_iso_recv_release_packets);
1308 EXPORT_SYMBOL(hpsb_iso_n_ready);
1309 EXPORT_SYMBOL(hpsb_iso_packet_sent);
1310 EXPORT_SYMBOL(hpsb_iso_packet_received);
1311 EXPORT_SYMBOL(hpsb_iso_wake);
1312 EXPORT_SYMBOL(hpsb_iso_recv_flush);
1315 EXPORT_SYMBOL(csr1212_new_directory);
1316 EXPORT_SYMBOL(csr1212_attach_keyval_to_directory);
1317 EXPORT_SYMBOL(csr1212_detach_keyval_from_directory);
1318 EXPORT_SYMBOL(csr1212_release_keyval);
1319 EXPORT_SYMBOL(csr1212_read);
1320 EXPORT_SYMBOL(csr1212_parse_keyval);
1321 EXPORT_SYMBOL(_csr1212_read_keyval);
1322 EXPORT_SYMBOL(_csr1212_destroy_keyval);
1323 #ifdef CONFIG_IEEE1394_EXPORT_FULL_API
1324 EXPORT_SYMBOL(csr1212_create_csr);
1325 EXPORT_SYMBOL(csr1212_init_local_csr);
1326 EXPORT_SYMBOL(csr1212_new_immediate);
1327 EXPORT_SYMBOL(csr1212_associate_keyval);
1328 EXPORT_SYMBOL(csr1212_new_string_descriptor_leaf);
1329 EXPORT_SYMBOL(csr1212_destroy_csr);
1330 EXPORT_SYMBOL(csr1212_generate_csr_image);
1331 EXPORT_SYMBOL(csr1212_parse_csr);