mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-18 18:23:53 +08:00
1655fc2e12
Make sure declarations stay in sync with definitions by keeping all extern declarations in common .h files. Signed-off-by: Hoang-Nam Nguyen <hnguyen@de.ibm.com> Signed-off-by: Roland Dreier <rolandd@cisco.com>
290 lines
8.5 KiB
C
290 lines
8.5 KiB
C
/*
|
|
* IBM eServer eHCA Infiniband device driver for Linux on POWER
|
|
*
|
|
* internal queue handling
|
|
*
|
|
* Authors: Waleri Fomin <fomin@de.ibm.com>
|
|
* Reinhard Ernst <rernst@de.ibm.com>
|
|
* Christoph Raisch <raisch@de.ibm.com>
|
|
*
|
|
* Copyright (c) 2005 IBM Corporation
|
|
*
|
|
* All rights reserved.
|
|
*
|
|
* This source code is distributed under a dual license of GPL v2.0 and OpenIB
|
|
* BSD.
|
|
*
|
|
* OpenIB BSD License
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* Redistributions of source code must retain the above copyright notice, this
|
|
* list of conditions and the following disclaimer.
|
|
*
|
|
* Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
|
|
* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
|
|
* IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#ifndef __IPZ_PT_FN_H__
|
|
#define __IPZ_PT_FN_H__
|
|
|
|
#define EHCA_PAGESHIFT 12
|
|
#define EHCA_PAGESIZE 4096UL
|
|
#define EHCA_PAGEMASK (~(EHCA_PAGESIZE-1))
|
|
#define EHCA_PT_ENTRIES 512UL
|
|
|
|
#include "ehca_tools.h"
|
|
#include "ehca_qes.h"
|
|
|
|
struct ehca_pd;
|
|
struct ipz_small_queue_page;
|
|
|
|
extern struct kmem_cache *small_qp_cache;
|
|
|
|
/* struct generic ehca page */
|
|
struct ipz_page {
|
|
u8 entries[EHCA_PAGESIZE];
|
|
};
|
|
|
|
#define IPZ_SPAGE_PER_KPAGE (PAGE_SIZE / 512)
|
|
|
|
struct ipz_small_queue_page {
|
|
unsigned long page;
|
|
unsigned long bitmap[IPZ_SPAGE_PER_KPAGE / BITS_PER_LONG];
|
|
int fill;
|
|
void *mapped_addr;
|
|
u32 mmap_count;
|
|
struct list_head list;
|
|
};
|
|
|
|
/* struct generic queue in linux kernel virtual memory (kv) */
|
|
struct ipz_queue {
|
|
u64 current_q_offset; /* current queue entry */
|
|
|
|
struct ipz_page **queue_pages; /* array of pages belonging to queue */
|
|
u32 qe_size; /* queue entry size */
|
|
u32 act_nr_of_sg;
|
|
u32 queue_length; /* queue length allocated in bytes */
|
|
u32 pagesize;
|
|
u32 toggle_state; /* toggle flag - per page */
|
|
u32 offset; /* save offset within page for small_qp */
|
|
struct ipz_small_queue_page *small_page;
|
|
};
|
|
|
|
/*
|
|
* return current Queue Entry for a certain q_offset
|
|
* returns address (kv) of Queue Entry
|
|
*/
|
|
static inline void *ipz_qeit_calc(struct ipz_queue *queue, u64 q_offset)
|
|
{
|
|
struct ipz_page *current_page;
|
|
if (q_offset >= queue->queue_length)
|
|
return NULL;
|
|
current_page = (queue->queue_pages)[q_offset >> EHCA_PAGESHIFT];
|
|
return ¤t_page->entries[q_offset & (EHCA_PAGESIZE - 1)];
|
|
}
|
|
|
|
/*
|
|
* return current Queue Entry
|
|
* returns address (kv) of Queue Entry
|
|
*/
|
|
static inline void *ipz_qeit_get(struct ipz_queue *queue)
|
|
{
|
|
return ipz_qeit_calc(queue, queue->current_q_offset);
|
|
}
|
|
|
|
/*
|
|
* return current Queue Page , increment Queue Page iterator from
|
|
* page to page in struct ipz_queue, last increment will return 0! and
|
|
* NOT wrap
|
|
* returns address (kv) of Queue Page
|
|
* warning don't use in parallel with ipz_QE_get_inc()
|
|
*/
|
|
void *ipz_qpageit_get_inc(struct ipz_queue *queue);
|
|
|
|
/*
|
|
* return current Queue Entry, increment Queue Entry iterator by one
|
|
* step in struct ipz_queue, will wrap in ringbuffer
|
|
* returns address (kv) of Queue Entry BEFORE increment
|
|
* warning don't use in parallel with ipz_qpageit_get_inc()
|
|
*/
|
|
static inline void *ipz_qeit_get_inc(struct ipz_queue *queue)
|
|
{
|
|
void *ret = ipz_qeit_get(queue);
|
|
queue->current_q_offset += queue->qe_size;
|
|
if (queue->current_q_offset >= queue->queue_length) {
|
|
queue->current_q_offset = 0;
|
|
/* toggle the valid flag */
|
|
queue->toggle_state = (~queue->toggle_state) & 1;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* return a bool indicating whether current Queue Entry is valid
|
|
*/
|
|
static inline int ipz_qeit_is_valid(struct ipz_queue *queue)
|
|
{
|
|
struct ehca_cqe *cqe = ipz_qeit_get(queue);
|
|
return ((cqe->cqe_flags >> 7) == (queue->toggle_state & 1));
|
|
}
|
|
|
|
/*
|
|
* return current Queue Entry, increment Queue Entry iterator by one
|
|
* step in struct ipz_queue, will wrap in ringbuffer
|
|
* returns address (kv) of Queue Entry BEFORE increment
|
|
* returns 0 and does not increment, if wrong valid state
|
|
* warning don't use in parallel with ipz_qpageit_get_inc()
|
|
*/
|
|
static inline void *ipz_qeit_get_inc_valid(struct ipz_queue *queue)
|
|
{
|
|
return ipz_qeit_is_valid(queue) ? ipz_qeit_get_inc(queue) : NULL;
|
|
}
|
|
|
|
/*
|
|
* returns and resets Queue Entry iterator
|
|
* returns address (kv) of first Queue Entry
|
|
*/
|
|
static inline void *ipz_qeit_reset(struct ipz_queue *queue)
|
|
{
|
|
queue->current_q_offset = 0;
|
|
return ipz_qeit_get(queue);
|
|
}
|
|
|
|
/*
|
|
* return the q_offset corresponding to an absolute address
|
|
*/
|
|
int ipz_queue_abs_to_offset(struct ipz_queue *queue, u64 addr, u64 *q_offset);
|
|
|
|
/*
|
|
* return the next queue offset. don't modify the queue.
|
|
*/
|
|
static inline u64 ipz_queue_advance_offset(struct ipz_queue *queue, u64 offset)
|
|
{
|
|
offset += queue->qe_size;
|
|
if (offset >= queue->queue_length) offset = 0;
|
|
return offset;
|
|
}
|
|
|
|
/* struct generic page table */
|
|
struct ipz_pt {
|
|
u64 entries[EHCA_PT_ENTRIES];
|
|
};
|
|
|
|
/* struct page table for a queue, only to be used in pf */
|
|
struct ipz_qpt {
|
|
/* queue page tables (kv), use u64 because we know the element length */
|
|
u64 *qpts;
|
|
u32 n_qpts;
|
|
u32 n_ptes; /* number of page table entries */
|
|
u64 *current_pte_addr;
|
|
};
|
|
|
|
/*
|
|
* constructor for a ipz_queue_t, placement new for ipz_queue_t,
|
|
* new for all dependent datastructors
|
|
* all QP Tables are the same
|
|
* flow:
|
|
* allocate+pin queue
|
|
* see ipz_qpt_ctor()
|
|
* returns true if ok, false if out of memory
|
|
*/
|
|
int ipz_queue_ctor(struct ehca_pd *pd, struct ipz_queue *queue,
|
|
const u32 nr_of_pages, const u32 pagesize,
|
|
const u32 qe_size, const u32 nr_of_sg,
|
|
int is_small);
|
|
|
|
/*
|
|
* destructor for a ipz_queue_t
|
|
* -# free queue
|
|
* see ipz_queue_ctor()
|
|
* returns true if ok, false if queue was NULL-ptr of free failed
|
|
*/
|
|
int ipz_queue_dtor(struct ehca_pd *pd, struct ipz_queue *queue);
|
|
|
|
/*
|
|
* constructor for a ipz_qpt_t,
|
|
* placement new for struct ipz_queue, new for all dependent datastructors
|
|
* all QP Tables are the same,
|
|
* flow:
|
|
* -# allocate+pin queue
|
|
* -# initialise ptcb
|
|
* -# allocate+pin PTs
|
|
* -# link PTs to a ring, according to HCA Arch, set bit62 id needed
|
|
* -# the ring must have room for exactly nr_of_PTEs
|
|
* see ipz_qpt_ctor()
|
|
*/
|
|
void ipz_qpt_ctor(struct ipz_qpt *qpt,
|
|
const u32 nr_of_qes,
|
|
const u32 pagesize,
|
|
const u32 qe_size,
|
|
const u8 lowbyte, const u8 toggle,
|
|
u32 * act_nr_of_QEs, u32 * act_nr_of_pages);
|
|
|
|
/*
|
|
* return current Queue Entry, increment Queue Entry iterator by one
|
|
* step in struct ipz_queue, will wrap in ringbuffer
|
|
* returns address (kv) of Queue Entry BEFORE increment
|
|
* warning don't use in parallel with ipz_qpageit_get_inc()
|
|
* warning unpredictable results may occur if steps>act_nr_of_queue_entries
|
|
* fix EQ page problems
|
|
*/
|
|
void *ipz_qeit_eq_get_inc(struct ipz_queue *queue);
|
|
|
|
/*
|
|
* return current Event Queue Entry, increment Queue Entry iterator
|
|
* by one step in struct ipz_queue if valid, will wrap in ringbuffer
|
|
* returns address (kv) of Queue Entry BEFORE increment
|
|
* returns 0 and does not increment, if wrong valid state
|
|
* warning don't use in parallel with ipz_queue_QPageit_get_inc()
|
|
* warning unpredictable results may occur if steps>act_nr_of_queue_entries
|
|
*/
|
|
static inline void *ipz_eqit_eq_get_inc_valid(struct ipz_queue *queue)
|
|
{
|
|
void *ret = ipz_qeit_get(queue);
|
|
u32 qe = *(u8 *)ret;
|
|
if ((qe >> 7) != (queue->toggle_state & 1))
|
|
return NULL;
|
|
ipz_qeit_eq_get_inc(queue); /* this is a good one */
|
|
return ret;
|
|
}
|
|
|
|
static inline void *ipz_eqit_eq_peek_valid(struct ipz_queue *queue)
|
|
{
|
|
void *ret = ipz_qeit_get(queue);
|
|
u32 qe = *(u8 *)ret;
|
|
if ((qe >> 7) != (queue->toggle_state & 1))
|
|
return NULL;
|
|
return ret;
|
|
}
|
|
|
|
/* returns address (GX) of first queue entry */
|
|
static inline u64 ipz_qpt_get_firstpage(struct ipz_qpt *qpt)
|
|
{
|
|
return be64_to_cpu(qpt->qpts[0]);
|
|
}
|
|
|
|
/* returns address (kv) of first page of queue page table */
|
|
static inline void *ipz_qpt_get_qpt(struct ipz_qpt *qpt)
|
|
{
|
|
return qpt->qpts;
|
|
}
|
|
|
|
#endif /* __IPZ_PT_FN_H__ */
|