2 * Copyright (c) 2009-2012 Mark D. Hill and David A. Wood
3 * Copyright (c) 2010-2012 Advanced Micro Devices, Inc.
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions are
8 * met: redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer;
10 * redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution;
13 * neither the name of the copyright holders nor the names of its
14 * contributors may be used to endorse or promote products derived from
15 * this software without specific prior written permission.
17 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
18 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
19 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
20 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
21 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
22 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
23 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
27 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30 machine(Directory, "Directory protocol")
31 : DirectoryMemory * directory,
32 MemoryControl * memBuffer,
33 int directory_latency = 12
36 MessageBuffer forwardFromDir, network="To", virtual_network="3", ordered="false", vnet_type="forward";
37 MessageBuffer responseFromDir, network="To", virtual_network="4", ordered="false", vnet_type="response";
38 MessageBuffer dmaResponseFromDir, network="To", virtual_network="1", ordered="true", vnet_type="response";
40 MessageBuffer requestToDir, network="From", virtual_network="2", ordered="true", vnet_type="request";
41 MessageBuffer dmaRequestToDir, network="From", virtual_network="0", ordered="true", vnet_type="request";
44 state_declaration(State, desc="Directory states", default="Directory_State_I") {
46 I, AccessPermission:Read_Write, desc="Invalid";
47 M, AccessPermission:Invalid, desc="Modified";
49 M_DRD, AccessPermission:Busy, desc="Blocked on an invalidation for a DMA read";
50 M_DWR, AccessPermission:Busy, desc="Blocked on an invalidation for a DMA write";
52 M_DWRI, AccessPermission:Busy, desc="Intermediate state M_DWR-->I";
53 M_DRDI, AccessPermission:Busy, desc="Intermediate state M_DRD-->I";
55 IM, AccessPermission:Busy, desc="Intermediate state I-->M";
56 MI, AccessPermission:Busy, desc="Intermediate state M-->I";
57 ID, AccessPermission:Busy, desc="Intermediate state for DMA_READ when in I";
58 ID_W, AccessPermission:Busy, desc="Intermediate state for DMA_WRITE when in I";
62 enumeration(Event, desc="Directory events") {
64 GETX, desc="A GETX arrives";
65 GETS, desc="A GETS arrives";
66 PUTX, desc="A PUTX arrives";
67 PUTX_NotOwner, desc="A PUTX arrives";
70 DMA_READ, desc="A DMA Read memory request";
71 DMA_WRITE, desc="A DMA Write memory request";
74 Memory_Data, desc="Fetched data from memory arrives";
75 Memory_Ack, desc="Writeback Ack from memory arrives";
81 structure(Entry, desc="...", interface="AbstractEntry") {
82 State DirectoryState, desc="Directory state";
83 DataBlock DataBlk, desc="data for the block";
84 NetDest Sharers, desc="Sharers for this block";
85 NetDest Owner, desc="Owner of this block";
88 // TBE entries for DMA requests
89 structure(TBE, desc="TBE entries for outstanding DMA requests") {
90 Address PhysicalAddress, desc="physical address";
91 State TBEState, desc="Transient State";
92 DataBlock DataBlk, desc="Data to be written (DMA write only)";
94 MachineID DmaRequestor, desc="DMA requestor";
97 structure(TBETable, external="yes") {
99 void allocate(Address);
100 void deallocate(Address);
101 bool isPresent(Address);
105 TBETable TBEs, template="<Directory_TBE>";
110 Entry getDirectoryEntry(Address addr), return_by_pointer="yes" {
111 Entry dir_entry := static_cast(Entry, "pointer", directory[addr]);
113 if (is_valid(dir_entry)) {
117 dir_entry := static_cast(Entry, "pointer",
118 directory.allocate(addr, new Entry));
122 State getState(TBE tbe, Address addr) {
125 } else if (directory.isPresent(addr)) {
126 return getDirectoryEntry(addr).DirectoryState;
132 void setState(TBE tbe, Address addr, State state) {
135 tbe.TBEState := state;
138 if (directory.isPresent(addr)) {
140 if (state == State:M) {
141 assert(getDirectoryEntry(addr).Owner.count() == 1);
142 assert(getDirectoryEntry(addr).Sharers.count() == 0);
145 getDirectoryEntry(addr).DirectoryState := state;
147 if (state == State:I) {
148 assert(getDirectoryEntry(addr).Owner.count() == 0);
149 assert(getDirectoryEntry(addr).Sharers.count() == 0);
150 directory.invalidateBlock(addr);
155 AccessPermission getAccessPermission(Address addr) {
156 TBE tbe := TBEs[addr];
158 return Directory_State_to_permission(tbe.TBEState);
161 if(directory.isPresent(addr)) {
162 return Directory_State_to_permission(getDirectoryEntry(addr).DirectoryState);
165 return AccessPermission:NotPresent;
168 void setAccessPermission(Address addr, State state) {
169 if (directory.isPresent(addr)) {
170 getDirectoryEntry(addr).changePermission(Directory_State_to_permission(state));
174 DataBlock getDataBlock(Address addr), return_by_ref="yes" {
175 return getDirectoryEntry(addr).DataBlk;
179 out_port(forwardNetwork_out, RequestMsg, forwardFromDir);
180 out_port(responseNetwork_out, ResponseMsg, responseFromDir);
181 out_port(requestQueue_out, ResponseMsg, requestToDir); // For recycling requests
182 out_port(dmaResponseNetwork_out, DMAResponseMsg, dmaResponseFromDir);
185 out_port(memQueue_out, MemoryMsg, memBuffer);
188 in_port(dmaRequestQueue_in, DMARequestMsg, dmaRequestToDir) {
189 if (dmaRequestQueue_in.isReady()) {
190 peek(dmaRequestQueue_in, DMARequestMsg) {
191 TBE tbe := TBEs[in_msg.LineAddress];
192 if (in_msg.Type == DMARequestType:READ) {
193 trigger(Event:DMA_READ, in_msg.LineAddress, tbe);
194 } else if (in_msg.Type == DMARequestType:WRITE) {
195 trigger(Event:DMA_WRITE, in_msg.LineAddress, tbe);
197 error("Invalid message");
203 in_port(requestQueue_in, RequestMsg, requestToDir) {
204 if (requestQueue_in.isReady()) {
205 peek(requestQueue_in, RequestMsg) {
206 TBE tbe := TBEs[in_msg.Address];
207 if (in_msg.Type == CoherenceRequestType:GETS) {
208 trigger(Event:GETS, in_msg.Address, tbe);
209 } else if (in_msg.Type == CoherenceRequestType:GETX) {
210 trigger(Event:GETX, in_msg.Address, tbe);
211 } else if (in_msg.Type == CoherenceRequestType:PUTX) {
212 if (getDirectoryEntry(in_msg.Address).Owner.isElement(in_msg.Requestor)) {
213 trigger(Event:PUTX, in_msg.Address, tbe);
215 trigger(Event:PUTX_NotOwner, in_msg.Address, tbe);
218 error("Invalid message");
225 // off-chip memory request/response is done
226 in_port(memQueue_in, MemoryMsg, memBuffer) {
227 if (memQueue_in.isReady()) {
228 peek(memQueue_in, MemoryMsg) {
229 TBE tbe := TBEs[in_msg.Address];
230 if (in_msg.Type == MemoryRequestType:MEMORY_READ) {
231 trigger(Event:Memory_Data, in_msg.Address, tbe);
232 } else if (in_msg.Type == MemoryRequestType:MEMORY_WB) {
233 trigger(Event:Memory_Ack, in_msg.Address, tbe);
235 DPRINTF(RubySlicc,"%s\n", in_msg.Type);
236 error("Invalid message");
244 action(a_sendWriteBackAck, "a", desc="Send writeback ack to requestor") {
245 peek(requestQueue_in, RequestMsg) {
246 enqueue(forwardNetwork_out, RequestMsg, latency=directory_latency) {
247 out_msg.Address := address;
248 out_msg.Type := CoherenceRequestType:WB_ACK;
249 out_msg.Requestor := in_msg.Requestor;
250 out_msg.Destination.add(in_msg.Requestor);
251 out_msg.MessageSize := MessageSizeType:Writeback_Control;
256 action(l_sendWriteBackAck, "la", desc="Send writeback ack to requestor") {
257 peek(memQueue_in, MemoryMsg) {
258 enqueue(forwardNetwork_out, RequestMsg, latency="1") {
259 out_msg.Address := address;
260 out_msg.Type := CoherenceRequestType:WB_ACK;
261 out_msg.Requestor := in_msg.OriginalRequestorMachId;
262 out_msg.Destination.add(in_msg.OriginalRequestorMachId);
263 out_msg.MessageSize := MessageSizeType:Writeback_Control;
268 action(b_sendWriteBackNack, "b", desc="Send writeback nack to requestor") {
269 peek(requestQueue_in, RequestMsg) {
270 enqueue(forwardNetwork_out, RequestMsg, latency=directory_latency) {
271 out_msg.Address := address;
272 out_msg.Type := CoherenceRequestType:WB_NACK;
273 out_msg.Requestor := in_msg.Requestor;
274 out_msg.Destination.add(in_msg.Requestor);
275 out_msg.MessageSize := MessageSizeType:Writeback_Control;
280 action(c_clearOwner, "c", desc="Clear the owner field") {
281 getDirectoryEntry(address).Owner.clear();
284 action(d_sendData, "d", desc="Send data to requestor") {
285 peek(memQueue_in, MemoryMsg) {
286 enqueue(responseNetwork_out, ResponseMsg, latency="1") {
287 out_msg.Address := address;
288 out_msg.Type := CoherenceResponseType:DATA;
289 out_msg.Sender := machineID;
290 out_msg.Destination.add(in_msg.OriginalRequestorMachId);
291 out_msg.DataBlk := in_msg.DataBlk;
292 out_msg.MessageSize := MessageSizeType:Response_Data;
297 action(dr_sendDMAData, "dr", desc="Send Data to DMA controller from directory") {
298 peek(memQueue_in, MemoryMsg) {
299 enqueue(dmaResponseNetwork_out, DMAResponseMsg, latency="1") {
300 assert(is_valid(tbe));
301 out_msg.PhysicalAddress := address;
302 out_msg.LineAddress := address;
303 out_msg.Type := DMAResponseType:DATA;
304 out_msg.DataBlk := in_msg.DataBlk; // we send the entire data block and rely on the dma controller to split it up if need be
305 out_msg.Destination.add(tbe.DmaRequestor);
306 out_msg.MessageSize := MessageSizeType:Response_Data;
313 action(drp_sendDMAData, "drp", desc="Send Data to DMA controller from incoming PUTX") {
314 peek(requestQueue_in, RequestMsg) {
315 enqueue(dmaResponseNetwork_out, DMAResponseMsg, latency="1") {
316 assert(is_valid(tbe));
317 out_msg.PhysicalAddress := address;
318 out_msg.LineAddress := address;
319 out_msg.Type := DMAResponseType:DATA;
320 out_msg.DataBlk := in_msg.DataBlk; // we send the entire data block and rely on the dma controller to split it up if need be
321 out_msg.Destination.add(tbe.DmaRequestor);
322 out_msg.MessageSize := MessageSizeType:Response_Data;
327 action(da_sendDMAAck, "da", desc="Send Ack to DMA controller") {
328 enqueue(dmaResponseNetwork_out, DMAResponseMsg, latency="1") {
329 assert(is_valid(tbe));
330 out_msg.PhysicalAddress := address;
331 out_msg.LineAddress := address;
332 out_msg.Type := DMAResponseType:ACK;
333 out_msg.Destination.add(tbe.DmaRequestor);
334 out_msg.MessageSize := MessageSizeType:Writeback_Control;
338 action(e_ownerIsRequestor, "e", desc="The owner is now the requestor") {
339 peek(requestQueue_in, RequestMsg) {
340 getDirectoryEntry(address).Owner.clear();
341 getDirectoryEntry(address).Owner.add(in_msg.Requestor);
345 action(f_forwardRequest, "f", desc="Forward request to owner") {
346 peek(requestQueue_in, RequestMsg) {
347 APPEND_TRANSITION_COMMENT("Own: ");
348 APPEND_TRANSITION_COMMENT(getDirectoryEntry(in_msg.Address).Owner);
349 APPEND_TRANSITION_COMMENT("Req: ");
350 APPEND_TRANSITION_COMMENT(in_msg.Requestor);
351 enqueue(forwardNetwork_out, RequestMsg, latency=directory_latency) {
352 out_msg.Address := address;
353 out_msg.Type := in_msg.Type;
354 out_msg.Requestor := in_msg.Requestor;
355 out_msg.Destination := getDirectoryEntry(in_msg.Address).Owner;
356 out_msg.MessageSize := MessageSizeType:Writeback_Control;
361 action(inv_sendCacheInvalidate, "inv", desc="Invalidate a cache block") {
362 peek(dmaRequestQueue_in, DMARequestMsg) {
363 enqueue(forwardNetwork_out, RequestMsg, latency=directory_latency) {
364 out_msg.Address := address;
365 out_msg.Type := CoherenceRequestType:INV;
366 out_msg.Requestor := machineID;
367 out_msg.Destination := getDirectoryEntry(in_msg.PhysicalAddress).Owner;
368 out_msg.MessageSize := MessageSizeType:Writeback_Control;
373 action(i_popIncomingRequestQueue, "i", desc="Pop incoming request queue") {
374 requestQueue_in.dequeue();
377 action(p_popIncomingDMARequestQueue, "p", desc="Pop incoming DMA queue") {
378 dmaRequestQueue_in.dequeue();
381 action(l_writeDataToMemory, "pl", desc="Write PUTX data to memory") {
382 peek(requestQueue_in, RequestMsg) {
383 // assert(in_msg.Dirty);
384 // assert(in_msg.MessageSize == MessageSizeType:Writeback_Data);
385 getDirectoryEntry(in_msg.Address).DataBlk := in_msg.DataBlk;
386 //getDirectoryEntry(in_msg.Address).DataBlk.copyPartial(in_msg.DataBlk, addressOffset(in_msg.Address), in_msg.Len);
390 action(dwt_writeDMADataFromTBE, "dwt", desc="DMA Write data to memory from TBE") {
391 assert(is_valid(tbe));
392 getDirectoryEntry(address).DataBlk.copyPartial(tbe.DataBlk, addressOffset(tbe.PhysicalAddress), tbe.Len);
395 action(v_allocateTBE, "v", desc="Allocate TBE") {
396 peek(dmaRequestQueue_in, DMARequestMsg) {
397 TBEs.allocate(address);
398 set_tbe(TBEs[address]);
399 tbe.DataBlk := in_msg.DataBlk;
400 tbe.PhysicalAddress := in_msg.PhysicalAddress;
401 tbe.Len := in_msg.Len;
402 tbe.DmaRequestor := in_msg.Requestor;
406 action(r_allocateTbeForDmaRead, "\r", desc="Allocate TBE for DMA Read") {
407 peek(dmaRequestQueue_in, DMARequestMsg) {
408 TBEs.allocate(address);
409 set_tbe(TBEs[address]);
410 tbe.DmaRequestor := in_msg.Requestor;
414 action(v_allocateTBEFromRequestNet, "\v", desc="Allocate TBE") {
415 peek(requestQueue_in, RequestMsg) {
416 TBEs.allocate(address);
417 set_tbe(TBEs[address]);
418 tbe.DataBlk := in_msg.DataBlk;
422 action(w_deallocateTBE, "w", desc="Deallocate TBE") {
423 TBEs.deallocate(address);
427 action(z_recycleRequestQueue, "z", desc="recycle request queue") {
428 requestQueue_in.recycle();
431 action(y_recycleDMARequestQueue, "y", desc="recycle dma request queue") {
432 dmaRequestQueue_in.recycle();
436 action(qf_queueMemoryFetchRequest, "qf", desc="Queue off-chip fetch request") {
437 peek(requestQueue_in, RequestMsg) {
438 enqueue(memQueue_out, MemoryMsg, latency="1") {
439 out_msg.Address := address;
440 out_msg.Type := MemoryRequestType:MEMORY_READ;
441 out_msg.Sender := machineID;
442 out_msg.OriginalRequestorMachId := in_msg.Requestor;
443 out_msg.MessageSize := in_msg.MessageSize;
444 out_msg.DataBlk := getDirectoryEntry(in_msg.Address).DataBlk;
445 DPRINTF(RubySlicc,"%s\n", out_msg);
450 action(qf_queueMemoryFetchRequestDMA, "qfd", desc="Queue off-chip fetch request") {
451 peek(dmaRequestQueue_in, DMARequestMsg) {
452 enqueue(memQueue_out, MemoryMsg, latency="1") {
453 out_msg.Address := address;
454 out_msg.Type := MemoryRequestType:MEMORY_READ;
455 out_msg.Sender := machineID;
456 //out_msg.OriginalRequestorMachId := machineID;
457 out_msg.MessageSize := in_msg.MessageSize;
458 out_msg.DataBlk := getDirectoryEntry(address).DataBlk;
459 DPRINTF(RubySlicc,"%s\n", out_msg);
464 action(qw_queueMemoryWBRequest_partial, "qwp", desc="Queue off-chip writeback request") {
465 peek(dmaRequestQueue_in, DMARequestMsg) {
466 enqueue(memQueue_out, MemoryMsg, latency="1") {
467 out_msg.Address := address;
468 out_msg.Type := MemoryRequestType:MEMORY_WB;
469 //out_msg.OriginalRequestorMachId := machineID;
470 //out_msg.DataBlk := in_msg.DataBlk;
471 out_msg.DataBlk.copyPartial(in_msg.DataBlk, addressOffset(in_msg.PhysicalAddress), in_msg.Len);
472 out_msg.MessageSize := in_msg.MessageSize;
473 //out_msg.Prefetch := in_msg.Prefetch;
475 DPRINTF(RubySlicc,"%s\n", out_msg);
480 action(qw_queueMemoryWBRequest_partialTBE, "qwt", desc="Queue off-chip writeback request") {
481 peek(requestQueue_in, RequestMsg) {
482 enqueue(memQueue_out, MemoryMsg, latency="1") {
483 assert(is_valid(tbe));
484 out_msg.Address := address;
485 out_msg.Type := MemoryRequestType:MEMORY_WB;
486 out_msg.OriginalRequestorMachId := in_msg.Requestor;
488 // out_msg.DataBlk := in_msg.DataBlk;
489 out_msg.DataBlk.copyPartial(tbe.DataBlk, addressOffset(tbe.PhysicalAddress), tbe.Len);
490 out_msg.MessageSize := in_msg.MessageSize;
491 //out_msg.Prefetch := in_msg.Prefetch;
493 DPRINTF(RubySlicc,"%s\n", out_msg);
500 action(l_queueMemoryWBRequest, "lq", desc="Write PUTX data to memory") {
501 peek(requestQueue_in, RequestMsg) {
502 enqueue(memQueue_out, MemoryMsg, latency="1") {
503 out_msg.Address := address;
504 out_msg.Type := MemoryRequestType:MEMORY_WB;
505 out_msg.Sender := machineID;
506 out_msg.OriginalRequestorMachId := in_msg.Requestor;
507 out_msg.DataBlk := in_msg.DataBlk;
508 out_msg.MessageSize := in_msg.MessageSize;
509 //out_msg.Prefetch := in_msg.Prefetch;
511 DPRINTF(RubySlicc,"%s\n", out_msg);
516 action(l_popMemQueue, "q", desc="Pop off-chip request queue") {
517 memQueue_in.dequeue();
520 action(w_writeDataToMemoryFromTBE, "\w", desc="Write date to directory memory from TBE") {
521 //getDirectoryEntry(address).DataBlk := TBEs[address].DataBlk;
522 assert(is_valid(tbe));
523 getDirectoryEntry(address).DataBlk.copyPartial(tbe.DataBlk,
524 addressOffset(tbe.PhysicalAddress),
531 transition({M_DRD, M_DWR, M_DWRI, M_DRDI}, GETX) {
532 z_recycleRequestQueue;
535 transition({IM, MI, ID, ID_W}, {GETX, GETS, PUTX, PUTX_NotOwner} ) {
536 z_recycleRequestQueue;
539 transition({IM, MI, ID, ID_W}, {DMA_READ, DMA_WRITE} ) {
540 y_recycleDMARequestQueue;
544 transition(I, GETX, IM) {
546 qf_queueMemoryFetchRequest;
548 i_popIncomingRequestQueue;
551 transition(IM, Memory_Data, M) {
553 //e_ownerIsRequestor;
558 transition(I, DMA_READ, ID) {
560 r_allocateTbeForDmaRead;
561 qf_queueMemoryFetchRequestDMA;
562 p_popIncomingDMARequestQueue;
565 transition(ID, Memory_Data, I) {
567 //p_popIncomingDMARequestQueue;
574 transition(I, DMA_WRITE, ID_W) {
576 qw_queueMemoryWBRequest_partial;
577 p_popIncomingDMARequestQueue;
580 transition(ID_W, Memory_Ack, I) {
581 dwt_writeDMADataFromTBE;
587 transition(M, DMA_READ, M_DRD) {
589 inv_sendCacheInvalidate;
590 p_popIncomingDMARequestQueue;
593 transition(M_DRD, PUTX, M_DRDI) {
597 l_queueMemoryWBRequest;
598 i_popIncomingRequestQueue;
601 transition(M_DRDI, Memory_Ack, I) {
608 transition(M, DMA_WRITE, M_DWR) {
610 inv_sendCacheInvalidate;
611 p_popIncomingDMARequestQueue;
614 transition(M_DWR, PUTX, M_DWRI) {
616 qw_queueMemoryWBRequest_partialTBE;
618 i_popIncomingRequestQueue;
621 transition(M_DWRI, Memory_Ack, I) {
622 w_writeDataToMemoryFromTBE;
629 transition(M, GETX, M) {
632 i_popIncomingRequestQueue;
635 transition(M, PUTX, MI) {
638 v_allocateTBEFromRequestNet;
639 l_queueMemoryWBRequest;
640 i_popIncomingRequestQueue;
643 transition(MI, Memory_Ack, I) {
644 w_writeDataToMemoryFromTBE;
650 transition(M, PUTX_NotOwner, M) {
652 i_popIncomingRequestQueue;
655 transition(I, PUTX_NotOwner, I) {
657 i_popIncomingRequestQueue;