| | import os |
| | import torch |
| | from functools import cache, wraps |
| |
|
| | |
| |
|
| | |
| |
|
| | sdpa_slice_trigger_rate = float(os.environ.get('IPEX_SDPA_SLICE_TRIGGER_RATE', 1)) |
| | attention_slice_rate = float(os.environ.get('IPEX_ATTENTION_SLICE_RATE', 0.5)) |
| |
|
| | |
| | @cache |
| | def find_split_size(original_size, slice_block_size, slice_rate=2): |
| | split_size = original_size |
| | while True: |
| | if (split_size * slice_block_size) <= slice_rate and original_size % split_size == 0: |
| | return split_size |
| | split_size = split_size - 1 |
| | if split_size <= 1: |
| | return 1 |
| | return split_size |
| |
|
| |
|
| | |
| | @cache |
| | def find_sdpa_slice_sizes(query_shape, key_shape, query_element_size, slice_rate=2, trigger_rate=3): |
| | batch_size, attn_heads, query_len, _ = query_shape |
| | _, _, key_len, _ = key_shape |
| |
|
| | slice_batch_size = attn_heads * (query_len * key_len) * query_element_size / 1024 / 1024 / 1024 |
| |
|
| | split_batch_size = batch_size |
| | split_head_size = attn_heads |
| | split_query_size = query_len |
| |
|
| | do_batch_split = False |
| | do_head_split = False |
| | do_query_split = False |
| |
|
| | if batch_size * slice_batch_size >= trigger_rate: |
| | do_batch_split = True |
| | split_batch_size = find_split_size(batch_size, slice_batch_size, slice_rate=slice_rate) |
| |
|
| | if split_batch_size * slice_batch_size > slice_rate: |
| | slice_head_size = split_batch_size * (query_len * key_len) * query_element_size / 1024 / 1024 / 1024 |
| | do_head_split = True |
| | split_head_size = find_split_size(attn_heads, slice_head_size, slice_rate=slice_rate) |
| |
|
| | if split_head_size * slice_head_size > slice_rate: |
| | slice_query_size = split_batch_size * split_head_size * (key_len) * query_element_size / 1024 / 1024 / 1024 |
| | do_query_split = True |
| | split_query_size = find_split_size(query_len, slice_query_size, slice_rate=slice_rate) |
| |
|
| | return do_batch_split, do_head_split, do_query_split, split_batch_size, split_head_size, split_query_size |
| |
|
| |
|
| | original_scaled_dot_product_attention = torch.nn.functional.scaled_dot_product_attention |
| | @wraps(torch.nn.functional.scaled_dot_product_attention) |
| | def dynamic_scaled_dot_product_attention(query, key, value, attn_mask=None, dropout_p=0.0, is_causal=False, **kwargs): |
| | if query.device.type != "xpu": |
| | return original_scaled_dot_product_attention(query, key, value, attn_mask=attn_mask, dropout_p=dropout_p, is_causal=is_causal, **kwargs) |
| | is_unsqueezed = False |
| | if len(query.shape) == 3: |
| | query = query.unsqueeze(0) |
| | is_unsqueezed = True |
| | if len(key.shape) == 3: |
| | key = key.unsqueeze(0) |
| | if len(value.shape) == 3: |
| | value = value.unsqueeze(0) |
| | do_batch_split, do_head_split, do_query_split, split_batch_size, split_head_size, split_query_size = find_sdpa_slice_sizes(query.shape, key.shape, query.element_size(), slice_rate=attention_slice_rate, trigger_rate=sdpa_slice_trigger_rate) |
| |
|
| | |
| | if do_batch_split: |
| | batch_size, attn_heads, query_len, _ = query.shape |
| | _, _, _, head_dim = value.shape |
| | hidden_states = torch.zeros((batch_size, attn_heads, query_len, head_dim), device=query.device, dtype=query.dtype) |
| | if attn_mask is not None: |
| | attn_mask = attn_mask.expand((query.shape[0], query.shape[1], query.shape[2], key.shape[-2])) |
| | for ib in range(batch_size // split_batch_size): |
| | start_idx = ib * split_batch_size |
| | end_idx = (ib + 1) * split_batch_size |
| | if do_head_split: |
| | for ih in range(attn_heads // split_head_size): |
| | start_idx_h = ih * split_head_size |
| | end_idx_h = (ih + 1) * split_head_size |
| | if do_query_split: |
| | for iq in range(query_len // split_query_size): |
| | start_idx_q = iq * split_query_size |
| | end_idx_q = (iq + 1) * split_query_size |
| | hidden_states[start_idx:end_idx, start_idx_h:end_idx_h, start_idx_q:end_idx_q, :] = original_scaled_dot_product_attention( |
| | query[start_idx:end_idx, start_idx_h:end_idx_h, start_idx_q:end_idx_q, :], |
| | key[start_idx:end_idx, start_idx_h:end_idx_h, :, :], |
| | value[start_idx:end_idx, start_idx_h:end_idx_h, :, :], |
| | attn_mask=attn_mask[start_idx:end_idx, start_idx_h:end_idx_h, start_idx_q:end_idx_q, :] if attn_mask is not None else attn_mask, |
| | dropout_p=dropout_p, is_causal=is_causal, **kwargs |
| | ) |
| | else: |
| | hidden_states[start_idx:end_idx, start_idx_h:end_idx_h, :, :] = original_scaled_dot_product_attention( |
| | query[start_idx:end_idx, start_idx_h:end_idx_h, :, :], |
| | key[start_idx:end_idx, start_idx_h:end_idx_h, :, :], |
| | value[start_idx:end_idx, start_idx_h:end_idx_h, :, :], |
| | attn_mask=attn_mask[start_idx:end_idx, start_idx_h:end_idx_h, :, :] if attn_mask is not None else attn_mask, |
| | dropout_p=dropout_p, is_causal=is_causal, **kwargs |
| | ) |
| | else: |
| | hidden_states[start_idx:end_idx, :, :, :] = original_scaled_dot_product_attention( |
| | query[start_idx:end_idx, :, :, :], |
| | key[start_idx:end_idx, :, :, :], |
| | value[start_idx:end_idx, :, :, :], |
| | attn_mask=attn_mask[start_idx:end_idx, :, :, :] if attn_mask is not None else attn_mask, |
| | dropout_p=dropout_p, is_causal=is_causal, **kwargs |
| | ) |
| | torch.xpu.synchronize(query.device) |
| | else: |
| | hidden_states = original_scaled_dot_product_attention(query, key, value, attn_mask=attn_mask, dropout_p=dropout_p, is_causal=is_causal, **kwargs) |
| | if is_unsqueezed: |
| | hidden_states.squeeze(0) |
| | return hidden_states |
| |
|