ioat: convert to circ_buf
[deliverable/linux.git] / drivers / dma / ioat / dma_v2.h
CommitLineData
5cbafa65
DW
1/*
2 * Copyright(c) 2004 - 2009 Intel Corporation. All rights reserved.
3 *
4 * This program is free software; you can redistribute it and/or modify it
5 * under the terms of the GNU General Public License as published by the Free
6 * Software Foundation; either version 2 of the License, or (at your option)
7 * any later version.
8 *
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
12 * more details.
13 *
14 * You should have received a copy of the GNU General Public License along with
15 * this program; if not, write to the Free Software Foundation, Inc., 59
16 * Temple Place - Suite 330, Boston, MA 02111-1307, USA.
17 *
18 * The full GNU General Public License is included in this distribution in the
19 * file called COPYING.
20 */
21#ifndef IOATDMA_V2_H
22#define IOATDMA_V2_H
23
24#include <linux/dmaengine.h>
abb12dfd 25#include <linux/circ_buf.h>
5cbafa65
DW
26#include "dma.h"
27#include "hw.h"
28
29
30extern int ioat_pending_level;
bf40a686 31extern int ioat_ring_alloc_order;
5cbafa65
DW
32
33/*
34 * workaround for IOAT ver.3.0 null descriptor issue
35 * (channel returns error when size is 0)
36 */
37#define NULL_DESC_BUFFER_SIZE 1
38
39#define IOAT_MAX_ORDER 16
40#define ioat_get_alloc_order() \
41 (min(ioat_ring_alloc_order, IOAT_MAX_ORDER))
a309218a
DW
42#define ioat_get_max_alloc_order() \
43 (min(ioat_ring_max_alloc_order, IOAT_MAX_ORDER))
5cbafa65
DW
44
45/* struct ioat2_dma_chan - ioat v2 / v3 channel attributes
46 * @base: common ioat channel parameters
47 * @xfercap_log; log2 of channel max transfer length (for fast division)
48 * @head: allocated index
49 * @issued: hardware notification point
50 * @tail: cleanup index
5cbafa65
DW
51 * @dmacount: identical to 'head' except for occasionally resetting to zero
52 * @alloc_order: log2 of the number of allocated descriptors
53 * @ring: software ring buffer implementation of hardware ring
54 * @ring_lock: protects ring attributes
55 */
56struct ioat2_dma_chan {
57 struct ioat_chan_common base;
58 size_t xfercap_log;
59 u16 head;
60 u16 issued;
61 u16 tail;
62 u16 dmacount;
63 u16 alloc_order;
5cbafa65
DW
64 struct ioat_ring_ent **ring;
65 spinlock_t ring_lock;
66};
67
68static inline struct ioat2_dma_chan *to_ioat2_chan(struct dma_chan *c)
69{
70 struct ioat_chan_common *chan = to_chan_common(c);
71
72 return container_of(chan, struct ioat2_dma_chan, base);
73}
74
abb12dfd 75static inline u16 ioat2_ring_size(struct ioat2_dma_chan *ioat)
5cbafa65 76{
abb12dfd 77 return 1 << ioat->alloc_order;
5cbafa65
DW
78}
79
80/* count of descriptors in flight with the engine */
81static inline u16 ioat2_ring_active(struct ioat2_dma_chan *ioat)
82{
abb12dfd 83 return CIRC_CNT(ioat->head, ioat->tail, ioat2_ring_size(ioat));
5cbafa65
DW
84}
85
86/* count of descriptors pending submission to hardware */
87static inline u16 ioat2_ring_pending(struct ioat2_dma_chan *ioat)
88{
abb12dfd 89 return CIRC_CNT(ioat->head, ioat->issued, ioat2_ring_size(ioat));
5cbafa65
DW
90}
91
92static inline u16 ioat2_ring_space(struct ioat2_dma_chan *ioat)
93{
abb12dfd 94 return ioat2_ring_size(ioat) - ioat2_ring_active(ioat);
5cbafa65
DW
95}
96
97/* assumes caller already checked space */
98static inline u16 ioat2_desc_alloc(struct ioat2_dma_chan *ioat, u16 len)
99{
100 ioat->head += len;
101 return ioat->head - len;
102}
103
104static inline u16 ioat2_xferlen_to_descs(struct ioat2_dma_chan *ioat, size_t len)
105{
106 u16 num_descs = len >> ioat->xfercap_log;
107
108 num_descs += !!(len & ((1 << ioat->xfercap_log) - 1));
109 return num_descs;
110}
111
2aec048c
DW
112/**
113 * struct ioat_ring_ent - wrapper around hardware descriptor
114 * @hw: hardware DMA descriptor (for memcpy)
115 * @fill: hardware fill descriptor
116 * @xor: hardware xor descriptor
117 * @xor_ex: hardware xor extension descriptor
118 * @pq: hardware pq descriptor
119 * @pq_ex: hardware pq extension descriptor
120 * @pqu: hardware pq update descriptor
121 * @raw: hardware raw (un-typed) descriptor
122 * @txd: the generic software descriptor for all engines
123 * @len: total transaction length for unmap
b094ad3b 124 * @result: asynchronous result of validate operations
2aec048c
DW
125 * @id: identifier for debug
126 */
127
5cbafa65 128struct ioat_ring_ent {
2aec048c
DW
129 union {
130 struct ioat_dma_descriptor *hw;
131 struct ioat_fill_descriptor *fill;
132 struct ioat_xor_descriptor *xor;
133 struct ioat_xor_ext_descriptor *xor_ex;
134 struct ioat_pq_descriptor *pq;
135 struct ioat_pq_ext_descriptor *pq_ex;
136 struct ioat_pq_update_descriptor *pqu;
137 struct ioat_raw_descriptor *raw;
138 };
5cbafa65 139 size_t len;
162b96e6 140 struct dma_async_tx_descriptor txd;
b094ad3b 141 enum sum_check_flags *result;
6df9183a
DW
142 #ifdef DEBUG
143 int id;
144 #endif
5cbafa65
DW
145};
146
147static inline struct ioat_ring_ent *
148ioat2_get_ring_ent(struct ioat2_dma_chan *ioat, u16 idx)
149{
abb12dfd 150 return ioat->ring[idx & (ioat2_ring_size(ioat) - 1)];
5cbafa65
DW
151}
152
09c8a5b8
DW
153static inline void ioat2_set_chainaddr(struct ioat2_dma_chan *ioat, u64 addr)
154{
155 struct ioat_chan_common *chan = &ioat->base;
156
157 writel(addr & 0x00000000FFFFFFFF,
158 chan->reg_base + IOAT2_CHAINADDR_OFFSET_LOW);
159 writel(addr >> 32,
160 chan->reg_base + IOAT2_CHAINADDR_OFFSET_HIGH);
161}
162
345d8523
DW
163int __devinit ioat2_dma_probe(struct ioatdma_device *dev, int dca);
164int __devinit ioat3_dma_probe(struct ioatdma_device *dev, int dca);
165struct dca_provider * __devinit ioat2_dca_init(struct pci_dev *pdev, void __iomem *iobase);
166struct dca_provider * __devinit ioat3_dca_init(struct pci_dev *pdev, void __iomem *iobase);
bf40a686
DW
167int ioat2_alloc_and_lock(u16 *idx, struct ioat2_dma_chan *ioat, int num_descs);
168int ioat2_enumerate_channels(struct ioatdma_device *device);
169struct dma_async_tx_descriptor *
170ioat2_dma_prep_memcpy_lock(struct dma_chan *c, dma_addr_t dma_dest,
171 dma_addr_t dma_src, size_t len, unsigned long flags);
172void ioat2_issue_pending(struct dma_chan *chan);
173int ioat2_alloc_chan_resources(struct dma_chan *c);
174void ioat2_free_chan_resources(struct dma_chan *c);
bf40a686
DW
175void __ioat2_restart_chan(struct ioat2_dma_chan *ioat);
176bool reshape_ring(struct ioat2_dma_chan *ioat, int order);
b094ad3b 177void __ioat2_issue_pending(struct ioat2_dma_chan *ioat);
aa4d72ae 178void ioat2_cleanup_event(unsigned long data);
e3232714 179void ioat2_timer_event(unsigned long data);
a6d52d70
DW
180int ioat2_quiesce(struct ioat_chan_common *chan, unsigned long tmo);
181int ioat2_reset_sync(struct ioat_chan_common *chan, unsigned long tmo);
5669e31c 182extern struct kobj_type ioat2_ktype;
162b96e6 183extern struct kmem_cache *ioat2_cache;
5cbafa65 184#endif /* IOATDMA_V2_H */
This page took 0.061458 seconds and 5 git commands to generate.