2 * Copyright (c) 2003-2005 The Regents of The University of Michigan
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions are
7 * met: redistributions of source code must retain the above copyright
8 * notice, this list of conditions and the following disclaimer;
9 * redistributions in binary form must reproduce the above copyright
10 * notice, this list of conditions and the following disclaimer in the
11 * documentation and/or other materials provided with the distribution;
12 * neither the name of the copyright holders nor the names of its
13 * contributors may be used to endorse or promote products derived from
14 * this software without specific prior written permission.
16 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
17 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
18 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
19 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
20 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
21 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
22 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
26 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 * Authors: Erik Hallnor
33 * Definition of BaseCache functions.
36 #include "mem/cache/base_cache.hh"
38 #include "cpu/base.hh"
42 BaseCache::CachePort::CachePort(const std::string
&_name
, BaseCache
*_cache
,
44 : Port(_name
), cache(_cache
), isCpuSide(_isCpuSide
)
47 //Start ports at null if more than one is created we should panic
53 BaseCache::CachePort::recvStatusChange(Port::Status status
)
55 cache
->recvStatusChange(status
, isCpuSide
);
59 BaseCache::CachePort::getDeviceAddressRanges(AddrRangeList
&resp
,
62 cache
->getAddressRanges(resp
, snoop
, isCpuSide
);
66 BaseCache::CachePort::deviceBlockSize()
68 return cache
->getBlockSize();
72 BaseCache::CachePort::recvTiming(Packet
*pkt
)
74 if (pkt
->isRequest() && blocked
)
76 DPRINTF(Cache
,"Scheduling a retry while blocked\n");
80 return cache
->doTimingAccess(pkt
, this, isCpuSide
);
84 BaseCache::CachePort::recvAtomic(Packet
*pkt
)
86 return cache
->doAtomicAccess(pkt
, isCpuSide
);
90 BaseCache::CachePort::recvFunctional(Packet
*pkt
)
92 cache
->doFunctionalAccess(pkt
, isCpuSide
);
96 BaseCache::CachePort::recvRetry()
99 if (!drainList
.empty()) {
100 //We have some responses to drain first
102 while (result
&& !drainList
.empty()) {
103 result
= sendTiming(drainList
.front());
105 drainList
.pop_front();
110 pkt
= cache
->getPacket();
111 MSHR
* mshr
= (MSHR
*)pkt
->senderState
;
112 bool success
= sendTiming(pkt
);
113 DPRINTF(Cache
, "Address %x was %s in sending the timing request\n",
114 pkt
->getAddr(), success
? "succesful" : "unsuccesful");
115 cache
->sendResult(pkt
, mshr
, success
);
116 if (success
&& cache
->doMasterRequest())
118 //Still more to issue, rerequest in 1 cycle
120 BaseCache::CacheEvent
* reqCpu
= new BaseCache::CacheEvent(this);
121 reqCpu
->schedule(curTick
+ 1);
126 //pkt = cache->getCoherencePacket();
127 //We save the packet, no reordering on CSHRS
129 bool success
= sendTiming(pkt
);
130 if (success
&& cache
->doSlaveRequest())
132 //Still more to issue, rerequest in 1 cycle
134 BaseCache::CacheEvent
* reqCpu
= new BaseCache::CacheEvent(this);
135 reqCpu
->schedule(curTick
+ 1);
142 BaseCache::CachePort::setBlocked()
145 DPRINTF(Cache
, "Cache Blocking\n");
147 //Clear the retry flag
148 mustSendRetry
= false;
152 BaseCache::CachePort::clearBlocked()
155 DPRINTF(Cache
, "Cache Unblocking\n");
159 DPRINTF(Cache
, "Cache Sending Retry\n");
160 mustSendRetry
= false;
165 BaseCache::CacheEvent::CacheEvent(CachePort
*_cachePort
)
166 : Event(&mainEventQueue
, CPU_Tick_Pri
), cachePort(_cachePort
)
168 this->setFlags(AutoDelete
);
172 BaseCache::CacheEvent::CacheEvent(CachePort
*_cachePort
, Packet
*_pkt
)
173 : Event(&mainEventQueue
, CPU_Tick_Pri
), cachePort(_cachePort
), pkt(_pkt
)
175 this->setFlags(AutoDelete
);
179 BaseCache::CacheEvent::process()
183 if (!cachePort
->isCpuSide
)
186 pkt
= cachePort
->cache
->getPacket();
187 MSHR
* mshr
= (MSHR
*) pkt
->senderState
;
188 bool success
= cachePort
->sendTiming(pkt
);
189 DPRINTF(Cache
, "Address %x was %s in sending the timing request\n",
190 pkt
->getAddr(), success
? "succesful" : "unsuccesful");
191 cachePort
->cache
->sendResult(pkt
, mshr
, success
);
192 if (success
&& cachePort
->cache
->doMasterRequest())
194 //Still more to issue, rerequest in 1 cycle
196 this->schedule(curTick
+1);
202 pkt
= cachePort
->cache
->getCoherencePacket();
203 bool success
= cachePort
->sendTiming(pkt
);
205 //Need to send on a retry
206 cachePort
->cshrRetry
= pkt
;
208 else if (cachePort
->cache
->doSlaveRequest())
210 //Still more to issue, rerequest in 1 cycle
212 this->schedule(curTick
+1);
218 //Know the packet to send
219 if (pkt
->flags
& NACKED_LINE
)
220 pkt
->result
= Packet::Nacked
;
222 pkt
->result
= Packet::Success
;
223 pkt
->makeTimingResponse();
224 if (!cachePort
->drainList
.empty()) {
225 //Already blocked waiting for bus, just append
226 cachePort
->drainList
.push_back(pkt
);
228 else if (!cachePort
->sendTiming(pkt
)) {
229 //It failed, save it to list of drain events
230 cachePort
->drainList
.push_back(pkt
);
235 BaseCache::CacheEvent::description()
237 return "timing event\n";
241 BaseCache::getPort(const std::string
&if_name
, int idx
)
245 if(cpuSidePort
== NULL
)
246 cpuSidePort
= new CachePort(name() + "-cpu_side_port", this, true);
249 else if (if_name
== "functional")
251 if(cpuSidePort
== NULL
)
252 cpuSidePort
= new CachePort(name() + "-cpu_side_port", this, true);
255 else if (if_name
== "cpu_side")
257 if(cpuSidePort
== NULL
)
258 cpuSidePort
= new CachePort(name() + "-cpu_side_port", this, true);
261 else if (if_name
== "mem_side")
263 if (memSidePort
!= NULL
)
264 panic("Already have a mem side for this cache\n");
265 memSidePort
= new CachePort(name() + "-mem_side_port", this, false);
268 else panic("Port name %s unrecognized\n", if_name
);
274 if (!cpuSidePort
|| !memSidePort
)
275 panic("Cache not hooked up on both sides\n");
276 cpuSidePort
->sendStatusChange(Port::RangeChange
);
280 BaseCache::regStats()
282 Request
temp_req((Addr
) NULL
, 4, 0);
283 Packet::Command temp_cmd
= Packet::ReadReq
;
284 Packet
temp_pkt(&temp_req
, temp_cmd
, 0); //@todo FIx command strings so this isn't neccessary
285 temp_pkt
.allocate(); //Temp allocate, all need data
287 using namespace Stats
;
290 for (int access_idx
= 0; access_idx
< NUM_MEM_CMDS
; ++access_idx
) {
291 Packet::Command cmd
= (Packet::Command
)access_idx
;
292 const string
&cstr
= temp_pkt
.cmdIdxToString(cmd
);
295 .init(maxThreadsPerCPU
)
296 .name(name() + "." + cstr
+ "_hits")
297 .desc("number of " + cstr
+ " hits")
298 .flags(total
| nozero
| nonan
)
303 .name(name() + ".demand_hits")
304 .desc("number of demand (read+write) hits")
307 demandHits
= hits
[Packet::ReadReq
] + hits
[Packet::WriteReq
];
310 .name(name() + ".overall_hits")
311 .desc("number of overall hits")
314 overallHits
= demandHits
+ hits
[Packet::SoftPFReq
] + hits
[Packet::HardPFReq
]
315 + hits
[Packet::Writeback
];
318 for (int access_idx
= 0; access_idx
< NUM_MEM_CMDS
; ++access_idx
) {
319 Packet::Command cmd
= (Packet::Command
)access_idx
;
320 const string
&cstr
= temp_pkt
.cmdIdxToString(cmd
);
323 .init(maxThreadsPerCPU
)
324 .name(name() + "." + cstr
+ "_misses")
325 .desc("number of " + cstr
+ " misses")
326 .flags(total
| nozero
| nonan
)
331 .name(name() + ".demand_misses")
332 .desc("number of demand (read+write) misses")
335 demandMisses
= misses
[Packet::ReadReq
] + misses
[Packet::WriteReq
];
338 .name(name() + ".overall_misses")
339 .desc("number of overall misses")
342 overallMisses
= demandMisses
+ misses
[Packet::SoftPFReq
] +
343 misses
[Packet::HardPFReq
] + misses
[Packet::Writeback
];
345 // Miss latency statistics
346 for (int access_idx
= 0; access_idx
< NUM_MEM_CMDS
; ++access_idx
) {
347 Packet::Command cmd
= (Packet::Command
)access_idx
;
348 const string
&cstr
= temp_pkt
.cmdIdxToString(cmd
);
350 missLatency
[access_idx
]
351 .init(maxThreadsPerCPU
)
352 .name(name() + "." + cstr
+ "_miss_latency")
353 .desc("number of " + cstr
+ " miss cycles")
354 .flags(total
| nozero
| nonan
)
359 .name(name() + ".demand_miss_latency")
360 .desc("number of demand (read+write) miss cycles")
363 demandMissLatency
= missLatency
[Packet::ReadReq
] + missLatency
[Packet::WriteReq
];
366 .name(name() + ".overall_miss_latency")
367 .desc("number of overall miss cycles")
370 overallMissLatency
= demandMissLatency
+ missLatency
[Packet::SoftPFReq
] +
371 missLatency
[Packet::HardPFReq
];
374 for (int access_idx
= 0; access_idx
< NUM_MEM_CMDS
; ++access_idx
) {
375 Packet::Command cmd
= (Packet::Command
)access_idx
;
376 const string
&cstr
= temp_pkt
.cmdIdxToString(cmd
);
379 .name(name() + "." + cstr
+ "_accesses")
380 .desc("number of " + cstr
+ " accesses(hits+misses)")
381 .flags(total
| nozero
| nonan
)
384 accesses
[access_idx
] = hits
[access_idx
] + misses
[access_idx
];
388 .name(name() + ".demand_accesses")
389 .desc("number of demand (read+write) accesses")
392 demandAccesses
= demandHits
+ demandMisses
;
395 .name(name() + ".overall_accesses")
396 .desc("number of overall (read+write) accesses")
399 overallAccesses
= overallHits
+ overallMisses
;
401 // miss rate formulas
402 for (int access_idx
= 0; access_idx
< NUM_MEM_CMDS
; ++access_idx
) {
403 Packet::Command cmd
= (Packet::Command
)access_idx
;
404 const string
&cstr
= temp_pkt
.cmdIdxToString(cmd
);
407 .name(name() + "." + cstr
+ "_miss_rate")
408 .desc("miss rate for " + cstr
+ " accesses")
409 .flags(total
| nozero
| nonan
)
412 missRate
[access_idx
] = misses
[access_idx
] / accesses
[access_idx
];
416 .name(name() + ".demand_miss_rate")
417 .desc("miss rate for demand accesses")
420 demandMissRate
= demandMisses
/ demandAccesses
;
423 .name(name() + ".overall_miss_rate")
424 .desc("miss rate for overall accesses")
427 overallMissRate
= overallMisses
/ overallAccesses
;
429 // miss latency formulas
430 for (int access_idx
= 0; access_idx
< NUM_MEM_CMDS
; ++access_idx
) {
431 Packet::Command cmd
= (Packet::Command
)access_idx
;
432 const string
&cstr
= temp_pkt
.cmdIdxToString(cmd
);
434 avgMissLatency
[access_idx
]
435 .name(name() + "." + cstr
+ "_avg_miss_latency")
436 .desc("average " + cstr
+ " miss latency")
437 .flags(total
| nozero
| nonan
)
440 avgMissLatency
[access_idx
] =
441 missLatency
[access_idx
] / misses
[access_idx
];
445 .name(name() + ".demand_avg_miss_latency")
446 .desc("average overall miss latency")
449 demandAvgMissLatency
= demandMissLatency
/ demandMisses
;
451 overallAvgMissLatency
452 .name(name() + ".overall_avg_miss_latency")
453 .desc("average overall miss latency")
456 overallAvgMissLatency
= overallMissLatency
/ overallMisses
;
458 blocked_cycles
.init(NUM_BLOCKED_CAUSES
);
460 .name(name() + ".blocked_cycles")
461 .desc("number of cycles access was blocked")
462 .subname(Blocked_NoMSHRs
, "no_mshrs")
463 .subname(Blocked_NoTargets
, "no_targets")
467 blocked_causes
.init(NUM_BLOCKED_CAUSES
);
469 .name(name() + ".blocked")
470 .desc("number of cycles access was blocked")
471 .subname(Blocked_NoMSHRs
, "no_mshrs")
472 .subname(Blocked_NoTargets
, "no_targets")
476 .name(name() + ".avg_blocked_cycles")
477 .desc("average number of cycles each access was blocked")
478 .subname(Blocked_NoMSHRs
, "no_mshrs")
479 .subname(Blocked_NoTargets
, "no_targets")
482 avg_blocked
= blocked_cycles
/ blocked_causes
;
485 .name(name() + ".fast_writes")
486 .desc("number of fast writes performed")
490 .name(name() + ".cache_copies")
491 .desc("number of cache copies performed")