mem-cache: Fix missing header in associative set
[gem5.git] / src / mem / cache / noncoherent_cache.cc
1 /*
2 * Copyright (c) 2010-2018 ARM Limited
3 * All rights reserved.
4 *
5 * The license below extends only to copyright in the software and shall
6 * not be construed as granting a license to any other intellectual
7 * property including but not limited to intellectual property relating
8 * to a hardware implementation of the functionality of the software
9 * licensed hereunder. You may use the software subject to the license
10 * terms below provided that you ensure that this notice is replicated
11 * unmodified and in its entirety in all distributions of the software,
12 * modified or unmodified, in source code or in binary form.
13 *
14 * Copyright (c) 2002-2005 The Regents of The University of Michigan
15 * Copyright (c) 2010,2015 Advanced Micro Devices, Inc.
16 * All rights reserved.
17 *
18 * Redistribution and use in source and binary forms, with or without
19 * modification, are permitted provided that the following conditions are
20 * met: redistributions of source code must retain the above copyright
21 * notice, this list of conditions and the following disclaimer;
22 * redistributions in binary form must reproduce the above copyright
23 * notice, this list of conditions and the following disclaimer in the
24 * documentation and/or other materials provided with the distribution;
25 * neither the name of the copyright holders nor the names of its
26 * contributors may be used to endorse or promote products derived from
27 * this software without specific prior written permission.
28 *
29 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
30 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
31 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
32 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
33 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
34 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
35 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
36 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
37 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
38 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
39 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
40 *
41 * Authors: Erik Hallnor
42 * Dave Greene
43 * Nathan Binkert
44 * Steve Reinhardt
45 * Ron Dreslinski
46 * Andreas Sandberg
47 * Nikos Nikoleris
48 */
49
50 /**
51 * @file
52 * Cache definitions.
53 */
54
55 #include "mem/cache/noncoherent_cache.hh"
56
57 #include <cassert>
58
59 #include "base/logging.hh"
60 #include "base/trace.hh"
61 #include "base/types.hh"
62 #include "debug/Cache.hh"
63 #include "mem/cache/cache_blk.hh"
64 #include "mem/cache/mshr.hh"
65 #include "params/NoncoherentCache.hh"
66
67 NoncoherentCache::NoncoherentCache(const NoncoherentCacheParams *p)
68 : BaseCache(p, p->system->cacheLineSize())
69 {
70 }
71
72 void
73 NoncoherentCache::satisfyRequest(PacketPtr pkt, CacheBlk *blk, bool, bool)
74 {
75 // As this a non-coherent cache located below the point of
76 // coherency, we do not expect requests that are typically used to
77 // keep caches coherent (e.g., InvalidateReq or UpdateReq).
78 assert(pkt->isRead() || pkt->isWrite());
79 BaseCache::satisfyRequest(pkt, blk);
80 }
81
82 bool
83 NoncoherentCache::access(PacketPtr pkt, CacheBlk *&blk, Cycles &lat,
84 PacketList &writebacks)
85 {
86 bool success = BaseCache::access(pkt, blk, lat, writebacks);
87
88 if (pkt->isWriteback() || pkt->cmd == MemCmd::WriteClean) {
89 assert(blk && blk->isValid());
90 // Writeback and WriteClean can allocate and fill even if the
91 // referenced block was not present or it was invalid. If that
92 // is the case, make sure that the new block is marked as
93 // writable
94 blk->status |= BlkWritable;
95 }
96
97 return success;
98 }
99
100 void
101 NoncoherentCache::doWritebacks(PacketList& writebacks, Tick forward_time)
102 {
103 while (!writebacks.empty()) {
104 PacketPtr wb_pkt = writebacks.front();
105 allocateWriteBuffer(wb_pkt, forward_time);
106 writebacks.pop_front();
107 }
108 }
109
110 void
111 NoncoherentCache::doWritebacksAtomic(PacketList& writebacks)
112 {
113 while (!writebacks.empty()) {
114 PacketPtr wb_pkt = writebacks.front();
115 memSidePort.sendAtomic(wb_pkt);
116 writebacks.pop_front();
117 delete wb_pkt;
118 }
119 }
120
121 void
122 NoncoherentCache::handleTimingReqMiss(PacketPtr pkt, CacheBlk *blk,
123 Tick forward_time, Tick request_time)
124 {
125 // miss
126 Addr blk_addr = pkt->getBlockAddr(blkSize);
127 MSHR *mshr = mshrQueue.findMatch(blk_addr, pkt->isSecure(), false);
128
129 // We can always write to a non coherent cache if the block is
130 // present and therefore if we have reached this point then the
131 // block should not be in the cache.
132 assert(mshr || !blk || !blk->isValid());
133
134 BaseCache::handleTimingReqMiss(pkt, mshr, blk, forward_time, request_time);
135 }
136
137 void
138 NoncoherentCache::recvTimingReq(PacketPtr pkt)
139 {
140 panic_if(pkt->cacheResponding(), "Should not see packets where cache "
141 "is responding");
142
143 panic_if(!(pkt->isRead() || pkt->isWrite()),
144 "Should only see read and writes at non-coherent cache\n");
145
146 BaseCache::recvTimingReq(pkt);
147 }
148
149 PacketPtr
150 NoncoherentCache::createMissPacket(PacketPtr cpu_pkt, CacheBlk *blk,
151 bool needs_writable,
152 bool is_whole_line_write) const
153 {
154 // We also fill for writebacks from the coherent caches above us,
155 // and they do not need responses
156 assert(cpu_pkt->needsResponse());
157
158 // A miss can happen only due to missing block
159 assert(!blk || !blk->isValid());
160
161 PacketPtr pkt = new Packet(cpu_pkt->req, MemCmd::ReadReq, blkSize);
162
163 // the packet should be block aligned
164 assert(pkt->getAddr() == pkt->getBlockAddr(blkSize));
165
166 pkt->allocate();
167 DPRINTF(Cache, "%s created %s from %s\n", __func__, pkt->print(),
168 cpu_pkt->print());
169 return pkt;
170 }
171
172
173 Cycles
174 NoncoherentCache::handleAtomicReqMiss(PacketPtr pkt, CacheBlk *&blk,
175 PacketList &writebacks)
176 {
177 PacketPtr bus_pkt = createMissPacket(pkt, blk, true,
178 pkt->isWholeLineWrite(blkSize));
179 DPRINTF(Cache, "Sending an atomic %s\n", bus_pkt->print());
180
181 Cycles latency = ticksToCycles(memSidePort.sendAtomic(bus_pkt));
182
183 assert(bus_pkt->isResponse());
184 // At the moment the only supported downstream requests we issue
185 // are ReadReq and therefore here we should only see the
186 // corresponding responses
187 assert(bus_pkt->isRead());
188 assert(pkt->cmd != MemCmd::UpgradeResp);
189 assert(!bus_pkt->isInvalidate());
190 assert(!bus_pkt->hasSharers());
191
192 // We are now dealing with the response handling
193 DPRINTF(Cache, "Receive response: %s\n", bus_pkt->print());
194
195 if (!bus_pkt->isError()) {
196 // Any reponse that does not have an error should be filling,
197 // afterall it is a read response
198 DPRINTF(Cache, "Block for addr %#llx being updated in Cache\n",
199 bus_pkt->getAddr());
200 blk = handleFill(bus_pkt, blk, writebacks, allocOnFill(bus_pkt->cmd));
201 assert(blk);
202 }
203 satisfyRequest(pkt, blk);
204
205 maintainClusivity(true, blk);
206
207 // Use the separate bus_pkt to generate response to pkt and
208 // then delete it.
209 if (!pkt->isWriteback() && pkt->cmd != MemCmd::WriteClean) {
210 assert(pkt->needsResponse());
211 pkt->makeAtomicResponse();
212 if (bus_pkt->isError()) {
213 pkt->copyError(bus_pkt);
214 }
215 }
216
217 delete bus_pkt;
218
219 return latency;
220 }
221
222 Tick
223 NoncoherentCache::recvAtomic(PacketPtr pkt)
224 {
225 panic_if(pkt->cacheResponding(), "Should not see packets where cache "
226 "is responding");
227
228 panic_if(!(pkt->isRead() || pkt->isWrite()),
229 "Should only see read and writes at non-coherent cache\n");
230
231 return BaseCache::recvAtomic(pkt);
232 }
233
234
235 void
236 NoncoherentCache::functionalAccess(PacketPtr pkt, bool from_cpu_side)
237 {
238 panic_if(!from_cpu_side, "Non-coherent cache received functional snoop"
239 " request\n");
240
241 BaseCache::functionalAccess(pkt, from_cpu_side);
242 }
243
244 void
245 NoncoherentCache::serviceMSHRTargets(MSHR *mshr, const PacketPtr pkt,
246 CacheBlk *blk)
247 {
248 // First offset for critical word first calculations
249 const int initial_offset = mshr->getTarget()->pkt->getOffset(blkSize);
250
251 MSHR::TargetList targets = mshr->extractServiceableTargets(pkt);
252 for (auto &target: targets) {
253 Packet *tgt_pkt = target.pkt;
254
255 switch (target.source) {
256 case MSHR::Target::FromCPU:
257 // handle deferred requests comming from a cache or core
258 // above
259
260 Tick completion_time;
261 // Here we charge on completion_time the delay of the xbar if the
262 // packet comes from it, charged on headerDelay.
263 completion_time = pkt->headerDelay;
264
265 satisfyRequest(tgt_pkt, blk);
266
267 // How many bytes past the first request is this one
268 int transfer_offset;
269 transfer_offset = tgt_pkt->getOffset(blkSize) - initial_offset;
270 if (transfer_offset < 0) {
271 transfer_offset += blkSize;
272 }
273 // If not critical word (offset) return payloadDelay.
274 // responseLatency is the latency of the return path
275 // from lower level caches/memory to an upper level cache or
276 // the core.
277 completion_time += clockEdge(responseLatency) +
278 (transfer_offset ? pkt->payloadDelay : 0);
279
280 assert(tgt_pkt->req->masterId() < system->maxMasters());
281 stats.cmdStats(tgt_pkt).missLatency[tgt_pkt->req->masterId()] +=
282 completion_time - target.recvTime;
283
284 tgt_pkt->makeTimingResponse();
285 if (pkt->isError())
286 tgt_pkt->copyError(pkt);
287
288 // Reset the bus additional time as it is now accounted for
289 tgt_pkt->headerDelay = tgt_pkt->payloadDelay = 0;
290 cpuSidePort.schedTimingResp(tgt_pkt, completion_time);
291 break;
292
293 case MSHR::Target::FromPrefetcher:
294 // handle deferred requests comming from a prefetcher
295 // attached to this cache
296 assert(tgt_pkt->cmd == MemCmd::HardPFReq);
297
298 if (blk)
299 blk->status |= BlkHWPrefetched;
300
301 // We have filled the block and the prefetcher does not
302 // require responses.
303 delete tgt_pkt;
304 break;
305
306 default:
307 // we should never see FromSnoop Targets as this is a
308 // non-coherent cache
309 panic("Illegal target->source enum %d\n", target.source);
310 }
311 }
312
313 // Reponses are filling and bring in writable blocks, therefore
314 // there should be no deferred targets and all the non-deferred
315 // targets are now serviced.
316 assert(mshr->getNumTargets() == 0);
317 }
318
319 void
320 NoncoherentCache::recvTimingResp(PacketPtr pkt)
321 {
322 assert(pkt->isResponse());
323 // At the moment the only supported downstream requests we issue
324 // are ReadReq and therefore here we should only see the
325 // corresponding responses
326 assert(pkt->isRead());
327 assert(pkt->cmd != MemCmd::UpgradeResp);
328 assert(!pkt->isInvalidate());
329 // This cache is non-coherent and any memories below are
330 // non-coherent too (non-coherent caches or the main memory),
331 // therefore the fetched block can be marked as writable.
332 assert(!pkt->hasSharers());
333
334 BaseCache::recvTimingResp(pkt);
335 }
336
337 PacketPtr
338 NoncoherentCache::evictBlock(CacheBlk *blk)
339 {
340 // A dirty block is always written back.
341
342 // A clean block can we written back, if we turned on writebacks
343 // for clean blocks. This could be useful if there is a cache
344 // below and we want to make sure the block is cached but if the
345 // memory below is the main memory WritebackCleans are
346 // unnecessary.
347
348 // If we clean writebacks are not enabled, we do not take any
349 // further action for evictions of clean blocks (i.e., CleanEvicts
350 // are unnecessary).
351 PacketPtr pkt = (blk->isDirty() || writebackClean) ?
352 writebackBlk(blk) : nullptr;
353
354 invalidateBlock(blk);
355
356 return pkt;
357 }
358
359 NoncoherentCache*
360 NoncoherentCacheParams::create()
361 {
362 assert(tags);
363 assert(replacement_policy);
364
365 return new NoncoherentCache(this);
366 }